var/home/core/zuul-output/0000755000175000017500000000000015133640323014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015133646532015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000242762615133646356020305 0ustar corecoreLoikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD 63I_翪|mvşo#oVݏKf+ovpZjl!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(_?u '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?~=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4ޙ-did˥]5]5᪩QJlyIPEQZȰ<'$VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:柺[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOO&Ack vz(cvb$^Nyo$@-t*"9BVQXodՔq[*ڔC"1Ȋ-R0ڱ.VhfFp佬)Wdڂ+ uR<$}Kr'ݔTW$md1"#mC_@:m PGEu&ݛȘPˬ-Ő\B`xr`"F'Iٺ*Dn@)y|tԞ!3Ír!S$,.:+dn̳BʺJ#SX*8ҁWח~>oFe-<uJQ|FZEP__'i(`0/ƍcv7go2G$ N%v$^^&Q 4AMbvvɀ1J{ڔ/hэKG9*W )IYO;Y4z⛢79"hK{BFEmBAΛ3>IO j uw/e{=t-3Pnef9[}=%G*9sX,¬xU&9'E&"/"cx}"mV5tŘ:cZ К G)]$mbXE ^ǭ8&>,0FЕ 6vAVKVCbr@25%L|v?33Iam:xg|Q'eű^\ʝCOU 6C],Ϣa!L9?ux޾8\0U'&2ihbvz= k ezƚO; -%M>AzzGݑT5sTkOt]﮶j=}Yf&OC"msht: rB<SYi&It1!ʐDN a$0Y&Hv]9Zzz+]}%b'$^LJ<\HGbIqܢcZW3{jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'7W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎd 5s gÁbNW\: "HK<bdY>OD d)VpA@A i"j<鮗;"(MC`ʶϊGi+39#gNZYE:Q۷9muB`9`LDhs4Ǩ9S`EkM{|D<˙ik; HD;;3!4 2Y.$Dwiu|+lO:k$]ԜYLUҞ6EmH:=yݫ_o{ʳwΛA ԀZk"f`.,ל{=wh,,z=%Rwóξ bnzFg;hwxDm/Ҳ mwKB d$tʸhi!{Q8AXT!Ov6dھ)]'pZrmbz3SX'\gya4bpnhb^Nf`ZT:tPt rG!I;rɳ Ӎ^hbO8.?IjRYОNnq[5N' ( oInf_nSI7{O;Brnm]{Lm`Mi Ctc}?0g`*钏xL(O>*3b\N7dYMcn`KݜȆqDDbZ:B#M^?tNG/\A.\pPO @:sg9_dTcxRk&%])ў}V\M6܎]Nbjgk`d]LGϸ.yҵVCL(lT@}d41ǽ2ҶQetwogזm'm.$]`6/}Op$Jbv7R0)pHOoc;un ]jU.mOm;>Qzi#f>夑3KմԔ萴%tx~~:7x}>yA>Z4Ӥ͋#+hI{hNZt 9`b˭`yD,Ȍ=6Z" 8L O)&OJ[m3Fo0"ZD|ijbɠM&HtpR:4Si גt&ngl9%islԃ)Hc`ebw|Ī3nZg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,l#u>cY^*xk%wCvpۊ ~;ɰ@ՙ.r{TM0:;}d8ۈ ݨW>Λ[VhiO̥̒_9$W!p}qz>xӾvs;mL?W+fx5Su{QqxST Ζ2%?T78{a{N8[lr`$pZdS=3jwlL Eڲ t|*Eq G.tݻgqW4ŠqBFf8 9Fk Afq#Owu5/>HB)iZBBy/@W,v&- 6CBjxk9ᤉ ,As~5YޜZ4ΓVYf'hkN7?qކ );B̗ߐu"8c`td 1xh˂U4U㳕Q`JRҴ225UY5li_;k+rlCf*8f- E_lq.JVh qC\/74Vւ'cJ߃cwL^obP) b5fWw1nrCj-di +92]5*#!#R^pn Ryw0/o|~8Ën]f);ٍVnSt%_>žBZ罉OMw0ŝl1!i~I%g{T~(?5P- j\hЪQxchKa?)]!Kqi4~Gvp!9V>^s9>NDͳ0*Vr'tdQu!4YhdqT nXeb|Ivż7>! {&~ĊL:}1*8&6f5 %>~R݄}WgѨ@OĹCuWapf٠r,$Zt 1-d}2Ozr@3?]^M ܓmBɽkQ| }^~n6Fg86}I h5&Xۘ8_->b ? z:9 Z>eUŎmxTL̵F8ՅX/>'q߾ӻoIQwdC#i  *!Jg,J?o(R~IBKb+{P.T! =ĦiTob d<>QHrSOKqWS7ѝBYǭ~TR"?^RNqٺQ '8{⼻wʹ`OW[F%mgXG2&XL_ʣ^"L/q2E3^-1ɢu^}G7X/q}^G)HDy=\]g|9i,p?ݼ\}e">|RJÌ=jO7ӑЇn-i3,1ֿ5GodgP\|bЯ݃_uM Wi·yT"^'~i6֬:v~m!8)"ۊB&{u:%lXZlvwohbL_#ǂsr_d >04SRm+0^PTi-"] O('@BKD6 {NmʐzRj.aQcb^CZ-uvpr CѐٱlGNzIveca=%1Qi F>wTLHUGӃ\sA֎Xpljlv ^tSȻ \cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ߉J%.Dl2ȱ%ܱ&6XƟ6qg(USok+Po$lwvmi8W_VT18V =| ub6QWCnY'"*aN08wuSEAVخ m3 o\` sHc# fqT .,ŀU|⦍߶/*~48âF,#[:y_YIpʼn)dk!J'Z5=r&; (y*b*O_ULT.ÔD[%s1,jЅ@k0Ցu֯dtKl$Y5O*GUڇvI`b0ο0~oI`b#FOf_$0!i rS/wvҍ%Eb/Ec|U9F-)L)ŘF`U:VK jeFrԋ7EDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃FZh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XN X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `5/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%~am.>!LcoJrKmqvez܅E9t6FZXgsreHhlٷ+ [}r:̓?W~e6>0E8`Jq-(ed;W¨:Ä&]䒿e;0:|$Ȃ1L-%;Ƅ{dɱL;V[bp>!n&աIJX1$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z< (1ZRÜ:OUM/vư{'jYXE4S/8 7: `/ +G\ U>]B2/n2=8) B gJ3bcKo̹ʇ\B~Is 2sO/I!}xV&\b<9$4Nve^آ]$LGF@LjKٕyzH 31Հm-XıUXF|\A-2) ' RG6h?āUŔyj[j_ӂ~ яA弆^bDyzǖQ8`jXbsK?l58,?YP5䜭ve9YFznTEf3Ja\,@2,?WYؾNr<V` =V[oB5!Z\ļǪЎr8@*ucѡv\[|s L-+y{5K@dzp`r"mũɸHNd"yc Pu>x2;W`_VR<aӗ&D<=h-Rר|/r _ǖھcߖ]G@Ն;UQG1 '3Jە Q88ASUȿ!:WѥLf21;d9OU᧯MR3V:<}xXh//T+coY5Ȧ4/m0NE(G2[+G~H'5ipӘ͏O +Px SPp.,?Uv]moȮ+F/wi m Nv(4uIKJcI1mj1r89C.x2Ua*o [GQ"h-Ń2R:HWi ƪ~q }Unn*1l;i4~=/yuU"񩇿cQo,FL?60͓jB/ `8`X%- d|]Rm]Rbt]Rh g-T4]] _ K/+KsB53csu52,M,|55nḵ|#=0< 3~uZM@n? }k0xT5%ϊj6 mW;g8+э˲몤+_^l?/*%:nw;<^hgw'U^d{A.oRo0|AzßPxA:ßKOvK89m%}| Ͳ/I_SS@j8852@S,e\Uξ凾o{h|2Ui_*-x*7(>ɹ|f:?TWuB/tB/TvaXSC\쵭̓m>:jԂR#A{8{7 濩ۺ؎'HrmV_c}`yc|?0uPur,HO+Q9Sl/lxI6~ /E&ώj #Q-_#y+Njz2 k^>Ơ>+c^ؠF,Qڣ2j:rŴ1b0UTTzj%/P(~d3-4뺆%Lt54Y(S ô@35-3L}ƨ-q}n:=_O&b}#]-,̋dZ^лX煸ðve2d|1Q kJ%Big(D NSɿQ9yPbR.: tI.š(c|-惀e: 1YVU^ Yr֌%o1U{#$̊Ic1R,RAZIlIb%߄yVY",&X_`:3c`!lb%䓰j/ Gi,MEiк\ػ(( qpQ0RL'/xOO` Œ]OZB[[Kx6h D6=UM-d3?|^9.׺}=͐ 2aD Ft,M`"q+U ~Ky;Ws w਍ q0/:ŕy6} ~mU,g<NV$(8Gqs0Ͳ|$]I;=~H˻sp(YW瓦dhI¢c9k9?=G=ŧ\4Hޟ@'7_{VWMbIe3iPGe|S_UQdrguq3$hSN㪨qMZ;iᙨ/OZ3 h.E4A&uޝǁcNA4PK-9& \\W{IRL/̾V,*Yhl|xo~?9wQɽXt^cc N xF؇KQy,*t )!;0فԅwS6qzSnhޟ8n<ݽϿ+mm yv^{[k=\}@/5^{9'sÊmyDhA$V(`Yn|7L†,DPv򦮲mzT4(K%A5+>GEU@M ,f;y]ō@jXLU8P T"Nq=@T٥H+-(qIs4 Z񑞨% S 9vJM|ohDaEӻx 2xW9}|4ˋ̟U]xV4Ȓ~" hPtM)W3IT!3"q4MU *7.#̽x2Y Gː FH݇T(_候cBj:LyЍS*Leʻudj2-iV%H^=e;ʚ-G,L)>!p}VHɒR,vXǴσUTfm톋qs˕ %̢u-aEhݽ9{#[r_f@Č=!yc}q<e}KoQ(3UO@Ig7tуej׊W4M0+|n<ؐCO/,H]bM k.Jz`9zTR> L`GyWD.UdhX:Ûڋ$G kWݓnJU:ީ)a vyv@0(\qU y,<[q<6xJ+" J`4tׅoC42_M*HIMe^̹l$]RTCr[b}b=C$ȉbbiDR%jmn(+s%)a7?RsƆ*>>!.C_#hpu2uI&S ۟fSMqeQM-"N6X*CK8kO7gIV^Dq3 })hF3d6:x%]KjtːY( *d`Џ:g\G}M[kES`?Wϕ9UNJ%5d̴s10`2gL㚾gUDvƛ!5dX>aigȢwYA&ZO`lI:(m#pnD3_h!;G,oF06\CF6 .ŐALJt'ݭvݢc6oaRѹtK&3U}OzmȖH̖[OIܩL(֣'ū&E w)Ʉwh;DkVե6"z1,mO;O^/4K}-GJ9f,;&h@d@3|UCS/gIv,z,O2>6y}LK2֙2:f|'t1-$w|QƓٌ&|M> @*Sh`ie1;K,Տq'QNq^g;w%v S!39m;r!&\Yd&wb$t2.^qMvaXWȗ"mذԳU5/H,z[?|KVq!lɼцXU딲ψv{YH5Wr(s\R.M/R,lq2Kv=]+̧r^9ڄ%ie L"ڿZka  1ƾiV]se۸%bsָ _,no(>m s {0ݘChE_j [s,mH5"}b!4@2֪ϮV/@?s\m0u\p0c EG|MX "0<{[[X?L ,Qj=5n s7d.B1RB7,Pw ] Kho bSBn a.sBO]C] g,D\w(mᣅnC;"ҠC\ @Ӯș102m4+MIuLTJ;4}@ 5VG Zs;5PzH o)]{,w:)@]_, i1h:}7~h,~uޮh&sGHlc -1BצlSQ^k^nqW-x/Fؘ@ɬ&"A:\A,ѩ:+J36VnL{<_;Oy^.3B@:Bv"L(J^%q745ص(_O?{Wܸ U/ϔd7xH*3HfF"HǺo7@R,{$1/3$@>X+S UYqYE<w%o1y}OQ8›0偫蘑UQDuvΪ_ϒI픇0:a?Y P:$f \&'#֔d~*ʪEo8)%'@a/x=] 0]hԕUaLr8E,V6x5e$ndV T|))5} 3۟Zc4`PV=DkFʷWc@<=ŽՖv~8Jwu 7կmQ'Ώf;N_Sן @e7!yBvٵ-pX =G@qib"vEDXt(*ݺ!~C!0SI-j_Bu0{h X6:kG0{x} 6YvZP>RܱcN!^(FaU7 Rp4x0Dd } >G5BPܞ͙MH% @PCu(f; \a|Sc'/ح~X!o!A]N7ߵP ݺ#oLW.HEӈ1ַ]ڏb!8q~;h?wW/sU|UH<Ե5~xߗ+ <9dlvٿֵv*8t*X2:B(]Sz΋G Fܞx^F7ܰvQWwuzcMߗSExa3ho<#X7amBXDamzs_4Y)v)%_%Spo{yuG%lW67D]ylԱ8ea_%/l*Yb_%D잸y Kҵ%z==E+^;6Yo|=}%OW/= |S qv+*;-OоY%kk |_1|틧a_ Um$J ]IwtN&W8w+t 첯tuZu˯sbJ䡾U;0Rȅ IKg{9 #.W3VHdj9L lG8hEp5'$aIBΠdr f˰W(S;q"T+}K $dhÛNѰJ;}Iz9G,jU]~,T&r^rzHWS"@asE9gRH Ѓ֠ĉd(/br«Mfi 6 P;z<ޒ8 I%U[gN-v@᚞e٣W jAT-*$MKXl7Q|-#l!H[rx@~1bD3Ih"#^(<.i].p-Up}Ӷcuhb(rž^,P3x#/~.@m撵'd=,AN6WPw'z>p3DzL|QI1qznXhj$̣13txq?o>2*@(X# ػdttd/gZbfo>h6[)"0'O ##pSl)^)vSK:I.>f]|Pg0Q$EIW$Zy纩ڃ]3o)\37O..ހ::Q׉P '-'6:k^<İ`'7Û8Mթ[y:@YP人o]1`áa>S3QeMq鲿qE"}pE@~= r+AJ%4|:dЯ!*U3LXѐYͭXf@{ԊUr ̮v}xJ1P~25|Ҩ,Y)xҀIkVuAŏӼYĞn1e&k0)y/׎ ||7 85-JO.~K8/  87bz7vI %/p8[]G@?'ӼGi: ρ^LܼUɟ9q|&~;>y T`6{蛯>:o_&*9AɇQðV˷O~Xd0qY֘#\ ((/8H-/Jto0=LhScnbxlkC{z8Q{5KyŀX|q53ǁC ף ]N0QaY=s<0rO(,侤< b@~/9x+i`70f\X&L~QQDL߆jWg';j )f<,ZޘGꚮipQߘ=3 F쉮>ޗpe=6HZX,N(Ag`mQ q5Fj?c-! $Ԉ(!< E|0X*lr*x5ue]Mz٣n*`<_66pֶؒ(`I+a3jfkk`I+<=p a`{o#j<9VMWʞ ]`x09z 39g9%aͮ;٭]ur~ ~^40C4@_d0`m, qD]W6,ë^dx[W vEm \ں&Tl/)BSC(k6 mN({lKB& )dL^8laen4 Kvdt6jTcKJ 5ۄjnN2B- 57 jjm@9$ڀPM/#ޒP{B6:꼌PgKB uۄnN2B- u7 km@9$ۀPM/#ߒP Kq)pvdmFk?6?Q.1y,3l$qL Y/eݒx_LT w`Ӂ1H&5KByfI˟[$`VeB~M?|<#5@{25m78 yr >zp ~DiTe_6].[ZrS8hd8+hd_<}?8'.yQ&1<)=t4%9WD[n>Z1X,C",įhH`Y 9.ҋ&%*Y G&ќB`(w"T#x$>?hZNmOrzrP I%gP ?-!nejϴVP6jrD*@|H|P'd.&0UWi eISoN.͇{lQPw,yn [L&ah J_2NINtt~YLn/bfz ;nu>xи s4bZ~IY_<-,8 X< a:DC̆a4vfI4 өuo _+e tk y/@ ĶN(|`K3*B2#Rɪ ip.^5 N}'Bb0Jw5kZxXn{bT~nap?)vLW[J<T%&غ5(üpJq^l^C~[,%-ʇ))p(c^xҀ3 ?nؠ dN6JFZI|HxRD"oկ&/۔PTOPROM)#Jv,ߜV.bǮ㏧=Wxr 8<qK\ ~'bKIq$³QJgYGCTq"t걀? "MMo@]"\Ie&g:&[@5q׷kFLr z[r'7_횲7`9nU9f\jH5f9!9VkՄxk`rk `cMWaI"FWixہMF/0D}K6&XnMnIOKQPl9×ѷէWV=7[/YuŅh#(NuU*Lb^kJJ◃ a*KtH6yR$i&yH*λ)Rt޿ 'TJģXV-R|n}cXXf4(bTQt`:gNWPe7ev6?Nq(y%c=Lh?Ӷ@ޡ9$_{_mXzcB˫yИx+gpOgybUjss_z>yK޵6z$׿ٕe}X5=@_%((i_j>L9/޵ wwB,u7H??_wCyZao/;lbcwNCŵyS?'`Ô2'<֨x#]77o)|D1i,m#ss"gOxSvHbտ\!7n;"%ؔId/K;_(fZcrޑ/ xwZ_ S3h#+~(. HIgK/Qࣘg*!0`4@ RNf}ZH >ÐK:b|/|X&]aq=:Gg5] 0 ~}!I* G#okfQ8f i t/ KjR!._(`Y!!p" fY). :Һ\8IX::m,8:1$v\jI# &ګ5݃ #w\#bj:˄~y0x'![ɖQ阾{'%o ^odS]ma7Qt\"jɽ#CʇdϏm5f,;f9}F'2w߲`m` )AI+4(t85Iٕi7&J&77 1hQN2K3 F'sc=mE4BC7 gt ?Y;q[ѣXȍ]GprAa4둬g{|eCr4]J|q#SG. 2;Ō /]8<+) )B/Г R( ܾm횅0l9c ̠G;߭"r 'r褳 ob voV&ГG~lA`8kKpA[z^zos(YL"=p*(9Oa2 C;wWU3hcM| ^h%/n{d'T]K·4=4~zo*ԼW9Abݝ<8? 1f7/~8˜0s`7n'0j3 ^89Q&B!X"jUݏͼj@7>9պ*lDς;8' Э90U#CU "!u|hXC;K ry$yyW8^yDf$E55- ^Gc(>GlX:f7rBOdF('е71 "aV*!Xۧjjw|Q+d=eФd(6Gʪ\ʾ{k(=ʷ|,۠K Lb3 5$#K{ȓKshSRGstbQJ.7N‚[k5i#TG*]i^ yڇ‚dh Ff"z8xϐ~JSDG1)17jsGJudlF4|["'zcqrԱ̼gvgm`;hY0|t'~l=t'Ln y}gP,<&2R8A)5LPVEsI/-_߰ZWYFZn??+(naAP]tSA~z?=뿴T6B2grbޘ`zbUÂd]Z`D0g,aKwc"=~ibil" 0fU4C HFqԍYmK-leSa %Wӓ+HF_S #$DIIϕ8G4x^BĕG3]]@fFŮukt\% YN h+J05702F ST8WRԴKttׯ%Ic?$iƏe x5y;3)Yw=醙,Zos[?YYd1b aㆬG@ԁm^Kj)`Xkܲ%yqG7⬚  /J^5 OM,^ _睹x_< o kL{P1KG™ϛP&Q{eN_mjr\01é@LHu&DMtǯ<qڎm2%:+0r D?fWD6,6۱;/$'O*q"Iī&N_`gXHER|;,yuw'KMPXzyJoI>NE^~&T/,i8k~H+oF0'Z9H fro>;5hG!89fUa@FLR,hz£5.Lf3?NNyoo dh ,i쌏]cF6B(c(c<$Jtknl>E!2pTGy^~yINm>:5Mzu7j['&_O XT(SG o gq';jwpVɍL A:AV1sd 4Rc^j79ݲ8]ez$sQOؠ5S]#m|\] nGg{xho4gtrrXx([|jV}1._~I*9>$g }\g<#~qSulKz)+GֽN,9wK99'/UR i#кhdtVQ 8"h#jrT,"zoh'' {ҡ6\&F_no:AcKGw{BaQ˳dF#sP>45yT$e+!QV3˞ e0R0b:dleZiE]eI=YI\)zNO%1[ǟAk%k +ϫ-_?n7i{%:m 8E45Ŝ1jI} F$S 1hmC%/: _Wy/Y Y?]EvKÂȂ"hMU εFP!iş|Q"c_9oǯ;鿚ǗTtވ2FB1)bm~*.ZM΂g*rB [G-VSK )' N-H1kςcpg0Pԧlr!FN.9p$F{ c kFWÇ6Kt,ɕ#JYXqԶ] vIm'ָ9ug tYNoI(S7,1ZwVERnc-['"J~)APu[ӪpJbfƔrb6q O Y;h:.YnmMvse#A\2M l˸]fC 'K`KWJ_!5mY,uj(gɉۛ 6Ih]YQٷ@+ ?rp)sBMNL^{-JpҜڦ7g |  ~~^U%M{8C yu0[ 5r\SzV1iN;yZf;ߩk ܇L:NenTtJt'MnLNqĻ'O@K^*.)<<@tUȱ캠N@S,׋>Jzyksvp-k[D?X<.{hmI)U$p6 p7EpOYM4\eѫii>`Kg9RN 7|sb7z;u*eπًOu =) Y~`Hj+|T}6 :R Fd) KڿYjr*ʒ3tՐB@ Ee?(/k'&/YAo7TޏEq` -dʺV;VE/PZdemTbCdSiP~F?6BjBE|8kiEb@&+?,)]KcLZNhV?ǎbS# P,=_*y4'LNK" GNwDqDEmH %mS-L.&BqT%htN3_vB|FknIgee9 :<տS5}k Luo^y擟O^?!o{os5x#<]G"%4xBp5-fLnsKy!|gu<3? 0EG[ Mb>AD0GGԅG'V`@ȣAUO}‡?5"z`_O&u( 1G1< MEv|ݠ&q ^pX1UCE1~ 0_b׉ #sn Dṉk_x3S]Y3 ] ۆC=0_Agp5h=12^kΗEeQ`Q8 czhp=G Yý}%1pjRvBY4W(3nC~[~9Do+ɕ ɠ_X;I#fp~0A7cZ7.8Eܺbz^zm )^wƎbsEEbLzFiߺ^|%znh}rwóx@.k,Y)2iY\~ka9% "`-wg3k N⛩MmLZ1"T[o.Nچue "'/ ŅF4D?lmbf7 ^FҬ Ĺ?Fߟ}ՠ3]B7U45A2>1zM`qin-hܽ㞉fĝon&gv4gsOSv}o9SL~LjNY$/zg"33g@-G B+ ԸI&rrg-)}cIw1͌IigrٺH{T)η)!+!$"ݱB!-iPwIݑZ'k$kǹ9! Z*-S!Rx)1Jn%Jh!iPvqh:1j;c~#VJh tUZytr`wKhuB̔Ej[ӻ[vuSD9IA;twJ2Qֈb,ypWCuG Բ[ХHTXτGrn 0>W 7/U=_TA1)XЎp[C 4׍nOxqlՉ1Fu^bYd=Et, `sR`)XK$$dДtf9Al5:U-'sfoo&e?իj{Rlh)SY]([G̎..:h*͟^_ ׯo{7OhQʦFC/V}z7 ]}f~9yQO +Jm0dVCdfM~iۓVۚs߽tF7 mp*i+WM8H^yFri tim7(ߟiὬxosx@& gu8\bwH)/),>59ǁ->,ku! E< lX3fMvV1xMI0*'DNpU ["bq=gyۢϋM, *#F`P5CµA}'彥AYG֒[D0^ǂ.Р/ʷXϫжk+}Kb>Dɽ_|_cU {dࠬN{o=FF_Ft'mKb&uk\| T7G8"[sϣ\y͍4 'wXr[%zG ?.Go#x떕[>a#3g4?|'\6z‚L`B6̂nh]H_EA) S'׭JK6 /JJҝs\{3{LY ^Li‡J/ƃܹ~ @% ß+ڻ@_Wm`U/,TM$Hu緙I뱶si!Zy|1"[K_!݇[#Ŝ,Z0#*7o܉a6Y8z,v}z+[Ċ$RO ŭfPB[? k )wz$X \cѢ/1HCx MKGEr x1a=~N_N}W_Q1<^']Apހ5I}6v U;s!Z'GA ɽ= v ]Jv˖Lں* (s)q^z, z) )BpDau5,]}D7VU117/?{`^`Ǣu:y ٻ # f p6SA4GG8uL|:{E)IO}Յ% GݼU*?p}Cq}\E}{_ѓwzP]Tc׻70L-s,йV )f gDsX/Ç]0"$FAg5=\ٵteZh>(9[a7D`#M=9^7{9>RR^j4B:e*1A dV{f1^b`;Dr5J#D`hڿѼX6k3U11ptYFT2fDZbl- ^aPES4w6 PNZm9˦]9YQC1tXL+D)4HbHPx#J4ͅ7H[@鋋wgfBlDR)i^x|4n4bA޲\v'Qe+w!IH' tزDd[A xZ:C,VPC\pr :ݜɛ7oKY˱m 8s5yP&;4Sn@:4ѝ)1܁C"6⭖U(V 3JĐc"1<(PăK5srl xn[a$ez)Z6eAxr+LДu|`s>T<' r' Nނ{&Ba̱ZMoO'?/q=+USv9bXoMY㓪ԕ\S'Y"ޯ މ ?h7 IS6-?] c끩lve8M;6h+K<+a?t#9SU> |(5|;&a]됩;#X:[\wG*#Ol#(ŭ}309y3@gw >nc[O r r-4gMX~CKevF@o4Q/4l:@J|%@"$3KON9F]0~Ġ0ϔVXI$Aɥ3KjOM`88@(̕e)ep &Kl1:CCD>@Q"&: ELմ+$X'WmF\i!p`Z_0/=XX$- ŒEeZ#2^eP&% J^AkE Yo %ڃRt@xhѥ𪤙rxr$hI )1Azm@4)-JxJ"kQqM 1܊,S+_~*:䧀\?9.W\-v#b[JY 9FA4.5 t)uJRLw|WQХĄkXd"HJ ~ hQZKR8>qxSű" =ZsQ  Kd8`Pa fq Q h)U oF9: $ 3Ρb 2,]l/FAFm`^KK$(f3:0Qל~`sKّوe [A t*Sui\֙"*!&(YR%6%*WJ,8V\hY8C!9FZ{i< )%@ci}wLKUgLs'؎:hOIq;TT!gH(V#8K |nrq L=9 qo Tu+;mxՇadi'92ݒw{9b$p λ-E|hS=3$GҐXIL_WUty MA81F AĨuqi'>BlK+@c.t|#ZND#+ΝEF)RC>: V$|h3TPUvT!p<3wn|*d|~;HN VZ=itϪ6u_Q1]WǕ;,J;L9.1]W1]8-:36h p>lU#"J9 D 'DNΠiAY4 !-Up<TyI3ȍy[q+L&DJFpJ-^H]$.i#x^xz589r<:h[$vҘs9CD, LB[|;PTGTB+LShlG$!p>y-(l4DDvg4ҩHP6(aD@2Eh9-Ks #/A&o 0JN{O|ZO;ipj Q2 -#}٦۴%œt4$MV-a2u>1QXҠ`8R*k<9H}` &; ,KQ zTHӹhXBʤgFKRPBeg3bo0-)͙_ʋ) pS;zNv?S\0TmvoJw9΃7̺[nǗ/ڡUؠ<39|Y,nFp1v4{Z+eA*vF#K CҘ YN 4CFwl E\/zr(-PJFHb6jn% ;l8v;󒯂n%DYt{]fL  8NrdME%yf?~پV.H+ ;+*=AgHqhTSh!hCm|]+ٌ;IeWQZZd9V2Mv,wD3JgV[0nڜ-)/uw^Eq!Lo nIahsoof&Oq^vzW :c'JՖS_ _?dȌ;fdϙ$@NKU@t𴁩^wi h^v ێh9PsFAFA6f'g@ (sQ 6n%0䕏R@>#w8%\! i T#, v3Z|3H:O9֊/Wq\|Ò7 Qa|h62'[񩄦>.s95~E"rHb5~c(0H᎙X8-$ݜMʚ^_G1?/*N ?"M'jAKM\\]h9׈d2sBȵSfus UD>ViNkɢo:<ۣ/܋p3bf*C̓AU*K$OmrYb0_,@e`hPe5O=%)c9cMRMׇz ̗KVK˷S1o%!s݇cUqsGh^M;{y;`44TՑU3i'ei_fZCzyŻ򹋷Xom[O/*}ݗңU'S7 !_=vn6;' T^KEPWhx,Ԯ/;:*-SѸٳ{\l|mO;`oɡâ. |}[8m߹Ao{5|jpX8EQU)_D?)-rk|{iݯ^A,uMfܫ8yרX~AEU#cQ$n\H'xG4hc>$'˽?jvΧF1'4/s|XbJׁiX>EURoeQ~zy 9[Jsɑkz`!W_Ov2}#Q(~7|j8E8rJW {PANs&aqV; rK*GP4jT(շR$?:VOH.)hb2]R=㍙u]GR(̴|hgow>4דy|٪8\%= ټ,G+; q h(^*~?ϗ&78aܣȾdv_oپ+ EzVkC檬qku#nHn[wu3 q_R%L[R]o$мjm-Co$a (9p#NT)@ka"9/ȆDc\N3A {.M;JvtLq,mΖp;Hɋ^?6OhMCTuOI*Pi 3FdlԊ2xϪ}p@85'YU5@ DW9+۬VvoҌɡo-ƶKIjewFmPӘopH*k  :c=!-Yk&d4^.K_1 })jMG>x0g~w;VNgO'1]MQcf+vmQ@=&T hB7UN/{S-b PP|V *FIv!CY3=wK(QJ+ёYf/{$_|{-o]YJ{IvZlϳ p4A6GY684)eǹ[̲a6/3,q:}}-~Es*9"Vv,þ~ 3 %m4/qk::_*^&_+ݟ(!^&:7Wsd<>(s"go71^wʬݥPd:pJRysH?if;Ǡ21t>m|7P4SND8ՏEއ]PoWYˁWe5EW@앗^Z(<ٸ<zCb7X\(K^5>WՏq7Ы\4惚Wk>$1v~5+& f&`?t-rkm S=_GTWE&\/kir7UɍK% c5M Dɫ\HJdltJ kSVvpؾ5f£7=K=[2¢~`q\vpݮԵCK_8A(zEL.z4etz'mX׿^̫g{WF| N.y>3 6_QYYϮJ,ͫ?N?ȧ^L W2|2eO--8y1`1j|]Q$^K;)gzҏ[?\ cq7YLWPr0`|n0 b})||(WBl)R(yb}dM<2W?X|Ç5moAV~;H?OC>\OS{J3.UN䷷jMϠHɄˏo;gy[hX οǛ+lԎzՃ*\˦a4 ^Y~8P;ݡf>ZW>]yG~USyUc{\=\¡4[jM:zk]kTv/w uUq}s߾! 4 a?{iӱhYm^dn~hC(-?/]OLcu}-ˉ>ͮ,rL뤽e "H͂ IR$-HqJkkPs7!Q3dm}㮏+^V?ui}(_p4YH2ur qsEѠ4I'8C rf讽)Cutd$E3Fl h(:&I ߷MA6RyZQOL$1Ԁ J!YGyKaMār x l9 nD# cU1T4scg<f&g F8Ѣh53D8_Q&U(h!`}-q:."p IDj1 q.mݘ(lF/4 / ʆ9OS }>sO6;kTA0p^Fp  A!qp]I1z[A DPK]ߥ[ (;G`6U&)˕6K v?%:L e$:T#Tv9oF69n6hÖ&/A6YPKgqedSiD)ϕRFGEK\` zD4%B&Mg)86>>DBZ#ɂ[V[hS4eM)|fC'vZOAKh#B-B#Ckǘ2D**]PCWj=iN<ʐSDYClb*G 5SNId6lKuAU@ Q4!#A8IO*D.ABmPDX(aj) +G`g@ы21[>PJpj9LQ2,n8J( C dȢ4*l C;>qtEjJ&8K {\(Z! $})<]p^(#eG6 CBpO8Tผh CH-pu@/zd^ Z`c@wmY4 ̮8èWZӤFl'=դzΝ nުu9Uunے*` m)"(zg4p\̧,,pWO +S )"$Є{1W*4 9bvH[4"Pa. :p,]," ˥trXD/@B]^j<`" P ˕3V40[ʆ4VH,@:ZR*nZñI#s5FaA{ŜE1N0ts\"0(l1IeS5I1T66% ~c5e `-2-7/TR2AV%TtgԌQ"Z")# VWjОU"x{x)(`~ZrABj2ZzC^vJ"F 02mDyYU+Dcé"j8(C e0>2a+"Ǜ ~\JΘ @<-ngge/F%Vy$' Wc@JRvN!&x?Mia;w3lٙ`YK.D49ac[-.D:l:G0^ BP8R`Vi>l:@G**Z@Ln]h)WVSc2(3Hv9뻭ՠƂBG1 $bDN W$2+B,h8P{ ȋ ":t@2ՎGM-2pd:A#,Y_~(yYQ"m dNI$ _O]ۖWCLm]Lܵ! EK0 w@܆upC$*f7]_GbT@mnk)jc2H =f n!*(S`b @pAs^ πB.ds3VU31b-Nm\(@ "uQ82Ggu]~cm\g 0,PM ;#RF W7 !ܠ0̹[c;gQ!ut",U@k69i LutОEwi`J fB޼թ'Q*gr]2+{уEX hߠtf&= H/6פf aC:kj:o8Hc[Mk=IS˴M¸qkֲ2%&P`ڢ[@7j!f=0p+Sݴ{OSRHuZtkHhS9)  r\O GC{o>(m>sWp0)A/7- 9py%\D[9DKhLwzTaH>"Kk=xP@ʭ`7z M҈  [W*-WEKH.: PQiQ0RXd'Z5uChep?{ڰ(RVGϊSZ]q}ݮL~Vw 58֠Yr=byP`ҧ"pՔq'k=[߾^rmn.>ƛZ#@K9Lת/Xf;`93#0²B3 _D2/ĺLtx/Kx>(5A=CPb.⌑TV;^@\0 ՗n CL*[.u _v)&d,¬"H"b Kq9,Xc.QuwEC@Tv""DoaL0Հ]GiZ$z~X;nVĦ:5{׀#n:o:rU駵Uk/Q'2]~>Xj._{<~2xiJggw Z{ĮIcvsedԇ\>+ІCh>ϼ*~_i> 3(s?=O5)ΟV Eblx#}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCbg>՗Z+?bވ}`^+#}}4>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb HFS͕l_>Z>J#HCbmb!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCbg>RaIo?|>0ʽԕj>R ؇>>KHCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$#[[vtӤmǗ׿<tN6 &ʁFT3y$X-iH9Aαn SEm_cJc<2~ UHlt:=\;w>SlitbhfcvNk>+~e>w,a`q ɬˮH#M2-ξZtTM@Ѽ$4;Z>CI[uZ5gO7/ rٯ[\?:94_&Y!Pp=W?hݖOh.~aj\;e]$U_DSbl樬ˣIe4c^2;'j=FX k?Hv\2X0kP5/F'C^E]do3 IK/|%hdV&9'@|=/Zg~ٸ wX'9yzX-ݴnjX߲FJi|mS;I(ewrS1N=P`V8ȭ~X͕x!n=!m u[T=5F%曨GkJD[.kpa}G8si_s3;g,?KO푻vjeI{yZN{67):pxdMUZ>,7' 4b>a/ T5NYcLf dQ<T={2x3s'p;O0dj7]H,ZB`޻)|mގV/֎>yѿ3oӛW/Fgs 8G7\4c))ޮޡZ G07\eZ&.g?.7Jlw4Un#_GdDD&C.ߡ:#cau\$/e }U/pJs7뀼Y_WϷtn v]BCʲݣzjt]L)p?T&tz29VG[ڵi7vPꞢ!wm7?v(Ճ[>*Nvp`5~)٧Iƫ༦1=0a'|-8w#QWvˇ2㺺<לThyw=Xb2Aucde>5<Փ+Gߚ;q3JЕg}%/21 ?O?Ϯߧpah&XƐe ǥ`Uf&Fd<~IgUD*".<כ}4[š,?ߜFJvn&+/hj+S{CmEKb' '&@[>Zlby&{M=v'`ꠒ7nT`!w4E剀}/%@xMPUGsۜP EhLүm.mOö0$twY57(}hT2:= wuPKh9=ov۷}{5̢fc=.\ۆJ`]B[@۵ξjcqϚjx$Dw?S8ӽyccc}S6|{kY>z1 vwnێ$ sT?PoV. z+VMZB{- )+,r]u.,~[F)]dLpp=faP K6<ȾG- fz?9#}V%酘kLyg'A\[=nHrR'c?ػ c_xyvרtLUI}wGAQ,BЀ`w$T y53BMpjJ2 -i0GN'~z,4X˕18IHx.E.୷Au\9ø8VS:>?5Wf4':Qr|2Zw=jaGg5$فuGkDz<~o+3'O»]#xh0k*]E&hy3v:^9s^_ff'?nyWN1c-IjdBB`毞Y\Pqw]RnNva5>όNluZ_#D`a;u[u.&onHѹ8${I0 }ֻ:sՃ1p}{$yѮcgk|y=u>Ź?{B{1cx,M/Y=^_b:Pz /W_ovo zV[]Tzܭo4$X7_sHbw$XxF$Sx ɹ֚2Ƹ!% 8% RN|5%4,h>M/է,7H ĩܕBBPH*%3X Fr6QCY.ƻ06&rЊI$prk|6%n3F<ݖ(lV]=M̹@גHEN b)K'j3G6cn>ͮUmټslXKKr" sv9x"G@<(LUjL.y\B15<[I 6V<ӡ1VQvjM[n믏 jwW8lJQC y?d:Hɋd`j>W-H!AJF-O\iypPƌE:pќlXC ~H0__bs@NO=M\6|~{ GG~{tƇš+?b?o~uVT Gt~#0cvLnZ/ Q PV@>; Z S,kG8,(^(`_"+F I=A6_7| ZfX6tQHQ:@jOWiKDQj | /uVtL=Ͷvj.a j^zmLHvCE*|i x7x93R4DIͲ$i{PG9E~d2qCl6B:〹[)Xqbh4' q^Ft_7eg4Y.ɡD"39>JO2g1]7thB |O ^`_Sm! 7 J­ʲ\8Oa P0|(]>z:=`]*3zCud;[x^$gs⮂h^,yr@F-uph8ܕj~)mamA("4DO34q/Jwb '-|(',$g[$l'R$Bb%WQA1_8+ |(x@4ɉh)&XO摶rM7vo.^kA FkAkLي P7u|r p{vpcuqv404p:Ex̛]֓\Gs]¿_:.ߤ_BQz5~Tq!8T0O'fLP` ^V^7mc.~?.Nyˇ^Բk+,XDr]ܖ>0@"# Jɫ3pH]bM1q-`QB" -"susZeIyeRғ. Mp CTKwNTyBsP] hV#F-p$R^Hm);62PLA%$9A|<g[>pZ9NW23dIˉD,qo yQOa@j:5o5h+W»KMaI4>?gDxͣgQyGcPp9"wzE/QVĬɴ:.=e UL+=I۰\(n2/W=*X[ h_pK6s3@!p}n(,ni{;h9n&,pىH1lv+.nc5j8 S T_R5o#tf[Jy;XXZdFL{C@ aX6MaiGsY1xaG+?.$ߤUfu}>Q7~+tWlF7,, pׅz{{ԪlcJP2͞RwI0O3m]z.tD| WGQ{ aNd!BquNT>?\4 G;Hrg&cy,!n{w7`1IPDd"jA [w8@;>`.WuE&_mֶ?+Vo7g Nomʄϵ˄QCY۲]7k:#Ϲ/1I TH]zUix܄Nfe{ttsuW'!<KyÎDhۙ(]O&}[%9=a%^z`N%E.y71nFjǝ>}0WgZZV} A"$mRw!} ^ "Jkl$ѐR"hJ˜[N>kyTP:174[Uv==sf "Tؑy~]2c}[Pzi|oOIYlu7QEg3l* ]:09<uҍo_VjEN A_ 2x|lݧU$3ΗPtdSv|TBl3直gf>P 2CZ1:]VcCPck.0UM.咍Gs>q^%t珦<IAc֡rTƌcJ::@B@P9q5I6d2;UO+}X`:R9eNv? |芌St~S}?^9UHA?0/D&HG |(˭UGLtJIIFտt=KUR˃ԧj ZO =:dԁX쉤JTdX:M G >NM_zhۘ*3gOޗ,9 }9þ:'|*u$ꮯ7툉d& $JSNY{;hLOdAtjQN):5G(2~ QEIqФi8aI?~ OM@ςH>8ZdiA6:uQmDm+> .sgtP3,{91;s\iNݫOh:V02” ?z=ty^+e 1F`\V}&c6n7)>mn4%. ]~:ZyLiN٦|:CAWIAC*KL+>D֞<)?!/;AhWܐ|l JT%A\{C*L7~ڶ|(k6y_jcG mP|*"f8Ó$PO8QR-&5T |bPQ -)EZX ,$VKM/v/| &)mSܼ ϤƮ2ƮM0 7kϪ z\#IE&Lڙ˴N]) G"HpƄ|9hM=^"G+iF]|46~=[m|\ wT )>z-&Ԉ]XLvW4!}28yξI&rlwj1i7AUQnߧY:gNA|O NGO&*z|t!C7SQ=̿QS\.Ÿ_Ugm_AWu12`ƀ.+g7,.!_&KE,c='']akG1^YҴO<@2s g1k%J\M iF4<|nx \ _1尜c VkrWgl-hvuxiZN_>hGqd9G0,gJGvN0VkQnC#VI1pk!m :kܠ8a Kc-8 >.>2FIզܼ־G`2X\EFO;Z֓ 8fA"1x#=č`;&cv #>ޅ/Stnͧ_槑Τk"u2mLWiU;†Th ,W|J'ms} aL-6-moq4'a0/z~n~l#V!`n4"!߼6J3\{J G|(E^"7={Ayb4&+fp\MM|z\͂\]hVY7 tVԂ3u31qjSrk=ukFcp=@2 </c a.Aep4&\g/}{Y <ͶI`XRaV}AO%咅ꡈsI'\m*W.`*$X`Fh;@CY265ڤ-/"IL )yS@mo-FdHSy^*xAaTM*MbiIbw,aZZ#bdycLL@uBǹlesGMIM*6K/MC]+x XFe4/v| 9Q"St3}]﫪GPP߯ RLO\2zptvQդ+Ŋ U]w@rX[UiL͆qM@J,r2mGBA\?FF^̐-laj>F|:sZX|nsI|v&=*cp6὚O^2y'F4[%+NeWd{,G"8YO,~ e81Rٕ{ 8,eKi50Ji|.f*,HugN5!-BIKw%\/afaFenQ`rJʘR߮/^hVov`6W۴&\j4)w:$£!HsHL_hnQspB@CYN˕u葄CÁ *ψO01i)Qi~O0 43i 螺R)hLAV~Ij ]M JN1[r@SN[5Ddcgt[wg8qݩ=9J r/ζ;+ϾO`b:&!:L(d]&(a7Qncu*I>GIV[6~Sz$#>_ז D1]ۤm4.$Wˆ1Ly )hUs:h\"J|[;NVy(}ۭM-*lբGbE62[;dZ`y%A{Pe??ZW5ME:rרc|v27\XI<,ׅHIף7HrIV<׵u7F dߤZIJ|ϼ^Tv=WvSVĴ1iu_[EvHDE/TW2>=E_BR&3]E>{ Ao!ņ B Nq i +{Ft;$vƳfn7p2ۈp]V.].sHY9a7]K&q3Q#AOPn'vhm>@rYgq]Y4t u&;C?d-Lc+m{-̃S&@)VL߽}>-wJ!j/F\ u|]&=S3{K%̎Q#eeΔ۬wϺ>}Ï̥UZr4V蠑9<9<=\*KB4fS:?`ɍsFA#/ssg=6:;Sw=Q U423>_V_o=dzKQ0USsӟCPhdfΜnS3ƞ!> i'زwR~^f#@Vdrsi&YYVzk5z#@5\f\{O7[ T5M#34}EU#?+Cf,Qh?E([2'ldg2륇1LjX-$KT'Syu;̜{T4Dgm*9B f΍$2[%t{W4233G#CVB9a# :hdfBD ~r@q>޴IlSPօEӥIVPtxfΛZ;Y~;Gv>_~Γ55̬QÝC~>m/73pn{ 5J=f1JIQꁧxT)O01 .g^@hz5"JҕJf(3"ZU3y 1 ]a*C&2"Q,Ft*jI&nhmhum6OV3 1oV/-rB%>HUt[ G?mf.hFyEۈAa:ƥ R%/]Em8H{=G^UTI!h ]0<:c@#3s(E)'<,'%^de`k$ JtQ$J1/'QᴮTz%QZ:$YILGZx%>HB(*2F(k9m9*xzY^w|YQ#pp*&_?G}bA(A]<^^`e `4v Ee*°6^~7 ;!)x140LS!sy&WoCTӝ Jtq$ e((9_oDGM wҴ`DGvHަd @eyzާzRI~2sr){w6[PqjQ ]Eu XF `/c[sHEdTu1 ڊ(OH/`^$  ~HԴbFyqBHosH8 &]X9õ msB7Cݎ(`@#WNzO FF `)/cS شAaMgY>'aapNV--fL! rw|yKLPåW`>B#3Eg ˍa h07SyҔ@Ff.%8]r.].` qJ 2+Cz sZcpL)ϓJ] 5Fj .h"0&y2$9twn }".^Fq z!/YG(1 ,ea/s^`frIɀ쇝/vk_vJ"?ƌO1cpL;1s5#u h\5jT)KJA`of޷ԒN5`D}7lA#Hd $]tZ&XqQ^Sӂ63Wc7t}N7#>/c ²` *]+pEz{HTreEۈw=HwN)R vL?IM;Kyݯ(Yۖmv4䲋4*bկP gSz|=8ު`64DFQ4pDbsf3~q,lO1D(uhA/C^ԯۄ~}MqFk1F(¯++]_6jeJ&|3YHK+!=j 'nz6):>K]_kDOߦ徥fAV֭80ku _,Nۏ/}%JOAee͡A=~rIࡗ=d Kfmxf^߼^PLS U;\?>[=+]νt S\%U|p N`h<8՛7@=_T]=9Q$\C#͛ucn{ԡJ]G }f C\,,֧*啦"$,5r l8˸ɬ R.j(%;F ͉6Z^xy]?aퟹS"Lֿ-i S_Ux%]Og6/$wl+'. /&_C'+㹞T}気nױߒsְm6f\~l$~ns!.ow{ s ^Iko/}q֖6nb_,n>mE2_qi\ŮCv>CBI}|֨Z[ ޸_?_OKQ6r5S{mfٕNN/T=A >0&\ϒ:y9Yee` %X?li'to7"|+mɋEYACWU,//.O1\f hoa=2Mzgl_\CL1 cČji ։hUI &gT7!ظn۰~}XRqq݈1qQq}kc-p +|} hC7BC5ڈ1FcU\pmvn>;78Xh>m(BB}A/<{ z%xd+>=ҩCB|JA!i.}x}-"qۮT17NnvFTb?l}N誥n/f?2&#j.AH )*jYdx#Y3/(e;DSD9Y-IyvcQr8it#H* .R\Z޹j90Ð<~E@|4m8GD鼬lVK m|D"zaQbU`i#qZќ[k.<q*+-BPR10z}“͚{{4rFş4,3Q8 p~$"cM(.@|.P9TN/]1ҵ6#@|IOaVsd CSln CjYG_?9QiD** ⡿S_)\#%FL\ ?CmF]pE#AF?ƲGc1\,[uH5ڡe\TaEX'VR8cEZ9;mjǡ%1B|T̨ %}k߅c7U8,,e"$nEƨ9lǴ$dMeb+UɴHeQ&]2-7&*OU-K_MZR܃h\=L$guZgR'Φyd8.~M:/i?Zb:s  aD/V2&hyXTYm:9{X() .^݈#*)*^TX8``W_pr;n̝Iil֊ŏyRJԙUC۵V45RA&CC{ |.h%PGw 1%RF"F_gպDVAr8zLP5A4a~1]Vނ*5V Z)47pJqÜ"\<58_(AE!_G@!d@Ƅ6,B\2K/R΅ؐdJ)S<Ĭ PeYvs~4* ZT@2_aVSǃ%klx 9ϼ JНéߝMn&5BR:u7p, *&`HNn|,*5Ŀٷ;p4SHi%9G>[<!T ʛpȚ-</<B Y1F%lFnDfTUܬ>.k8}H{"DǸ_?7qs/SҠ>LGg+U'N40>ICl[/EA3?_Xz&:Y$ĽCkp,H ⻯&mW ) **sd5$wkj-q:kwOl\v7̞*w~cx71ސ1m|hNI룄woqSy꿧uڋ bDmQ9Q=bqAg?2_7B.Ep|$]+[0`~K/]rX@|ͧ]C B ١A-b``FC=qX1|\XyYgL3Ӹp)0IQ־>}}!mV!gJx/(q/,O_C=qXQ_=ܣ'u|HpKvfy̷*q*nSEM>eU+(&ʇJ޺j.?ZmHCUm mIӷ6.EdrxYvZ.}Ȁ2[(%y #kaqRdpzE愉i[i[!ۖs\wvnSvN gi5 RvCVF uWNQ8K'Ap O*4N+AH[Dz<?=M/n5ueQY{.Ot c`'W_%NL̘Oa#I*"e# 8MsaGE%st̟V6̯  !f~-(s.BvE%u9$sEIt}'"h(5Zr!X)`,0nLJCzB|1n<pC}0o_o uta1V)NcbĮv*<:,($V3 U=cAu3uA! z!jb7 )3";$EVdbZ/u땇N_~<?#x!vp:zD@ l\]ˡFdZdΨwZ6CZrBpd8Zbf^Kj1C`]Au<1'8wt A3FCȋ<4CB:DzAAa4XA Ō)Sp0I2-e"K_ׅu@|{#1ss!U|08w*uV[\kcT. o+X1pY(se$KD^GNe5vƫE?I Ñ+@\XJ 髴,&^Bϲ9EEz B1hv~6D,Y +r2ΐsN҂]ݪz|v~ͰgX<{*Rou ;Xs{2c ZMUVHY]iӵ{j.]{i!>x`Vb &| =ͤDsvcfY]"GG#:(KE_$)jsĸIе6Wӹw{8Hw6vuSp!ćȌQAhw._ /PG?c,jyٹ9"X u9,UW! Yx^$;tۢXcf;eR y24{t A 'io *WT4U! vw4"hT*|Cy< NJ*wXͽ*Mc커M C\0Ԣ8ZY3siJQ{`p!H\IC2KGz=jjZLͱ fC=D|HCB6%?{-R e}R9IlwtU GӷiPG 11Ǒ&(Y |8t1-~%:i޶wmCS&>蹮FmKqfWF 7ͲHTefXo/UY'*oGRğ\3w.7F#*I3"(!E@Ę45:c! !-]!ħWQRL?1vF3Gmђ+C=ělcyQJny\~EI&|@UdrZAY{;;//yG?̊b =]kK]ד_~Y?YOmd`;}Jyd5/ɨzOߦ徥fݼae+\S* ["|xK~V9,.?n H9%^+g׳ 7~bȝ/X?g'e ;C ݞ;w j> ߰AM-ˎ+Y&UE272f5tm5E3L[9n`huynmO7jݯ+7Yba?JÎ~Ʉ 1 ˴"7YIZ"\ I)hKT{ȍa-p~ ,f'K .H/ %E3W˖ڦg ,~U,1Kt=3?|AW+ȱݚfogz= ũ^o7g|i2Ã՛]A-= .׍TOkíUPn ftv5^}\ZWإt?GifG5xfuyqpϺ-R>`~iX" 5;;};˶?tҡ-WR>7+/܆0G7I\vT_S;ҟvVzf^sj]< 4\`BH.s舣l:J`n}>3 oՉ JJ4R(r+[B/iZ "DhګwEo*?,[uWGZI?%[THn,}4=6`zWKÜ$cqkr@cS?g+SDKG1C8c~rgLwry:SmI\P@+9# ?Q8=O܇6v?yz }Xς.XХ=o{><ݞ/L!σ_KɄ2HlU <Vܷ{ 1+b~{UrަD˥Iu7!,[r-ϡBs8 B^Okd p{K}~ز%46\6hi +t K0qQ* 9kI#ML u_/z \ dEeUh|EU ZO6[1,8;'#v15)e^TS(S p 'Y}6w_ͧǿ͑ߛI<`Bսvg={Dbms?!\``dOΔd0bxFN ˓82hl-q;yGg#,)Jǽ x 2>.#Up kӋi0:[r 8 oߞ..!1SFRvͫ -eWa $;cZ=_\r4~}|m3T /pFólxQx-?5sGq ǸR)]6d$ԽdX!'YYNƣO+F74C3*S] jX5lMuSɄ?&3J|^甜>mǠQJՍ NA?o.}GB߁ p`Q6uq0Gt`4emz+LҲQ/^@6@=t7r~m[؇WOЋ88A>́AX} N>~Dիlz/4r i@):]nG~_=G𩎆MWbE 1p[eq934:2N\/1lfp|6D юYGjmAMDscU#T[!$㵡;3jz槷&wt\ol֖5d؞t̲`ymƒ$  LUi[2moV7XkˏIJlDPSP Q^4E]ʅ(.fJoNyu%YBi%YrsNO:{,Wr^&"> 3SrMY#jEeZ1wAl"&$~Θ ,xGxiwp&\u<֜߮F +8MeapsǃɍB`p+eM߽p/`C|{\vx6[g\+.wۗ+\ƧmՠkvD8~uA5)A5|2+VnVixFt.Y M 6AN8yN}UmWxΐgJ=>X֐Ȝ `hRxq퀸]; ͱ2iFJ6wvhm$of~2A-g"7ݬ-.nrK 0ͬ PL "5h vo_dA@ۘPNoocT3g2htվϮ3NAkʫOLD"O ht:w9.Pf( FyPT&FhtcZgO/9%7htgb84cȁո D̈́D,~Zsvs bb(\R0ޡ{mw *Z_.~z>:Y{ D"y &Zi FaxII-_7OLq-.37jS.3XB?$]!Bi bB 4:fG3 ʢt27뗢EvZ!ǫUZ\}Bv,Z)B- ~{m@),!|(Za-u %V+:5=-`kʀ?Ѱ|AIu*PkDkģGGir\뽙m3Ph^dRi $00/)fۨRb@c ќ^Ɛ{5TJu7Jtag|v Tc~TJ~#DP#b HCƇdTW 'L93 ABVᐕ9QAwӫ-3: JQ"sȐv [urW0: x1QJ R&y4V@G{) ,\!8+1fNf7bIcPP( 'bZ'sl\+pwRLA* Fia~j)zl\94DR؃Xp4dr_MiK'9szګ]xHXn7-x OaC: tndmZ3ty1шksN1gR51'|eg%,f XjcJ(ZQث9*c*rP 90HH";$/- @K@LZ {dV:fY'u {Ud4aɍE D L/Xtn!p 4=?ED>Q}Xt;#i?L0( Y^{F̡;:z a-8C@'Ypbż(NH P!q+6;PZ(@chlb (8r\ie!Pk@|M@b`aamDJ|T`w{YRbH7g|9 4:fST,J1([φ<=e4UJ"*V;BX<oehEE=ڇ/dT8A 4:f>hQam[(fβSifEApE`c0N㊊6!pRA<.6IJWҪȧYLJ9pe8`;DJ)v;[:gI.8 ҉Y䰠VpMV %ɳJs=sK`]#c=+Ŏ*JbCJ-ATI6(>L|> hFѸcS)Ho1s̉ -.;vͅF>4TY[JrBp VRm'oV3xvigq3XɊ1|%сj]q$`8yN|jmR/R{)}*3rpV 52FQĬKtwwtIcp gr"G+9NpN3! خ39csE9ʩ4mܶ[wKɯ,gghv? M%'0uk# ?$D6r*p} g6/SGk S܊E n듫r`XE&尼|aXNsC6sqeRӶ -K*LS o?Cf V,1JV\B")D(QNp @lr=H$Pvia-`tN 71;no٧iunsg  Ozo>P$>@c+u^vxFjcH,IE2%Mgj|Ƒ"xW\@q^wxboH*Ց](ȶmv4dr8fHΠ~2vxQ=dډWv6z.u蚰KBic#hjG1&$kach+6Mxzfhgd'ڝ9\JltHMц{x67/s4^:6.]D 5LK'8hbjy=3t,>ݹ|Zٙ[m 2*?*e"m>ze"P>n١AhD ==U"Ѡ`L,P5܃&Sƍ[ݞ3s؃`lu'B"pGzw l}`AdߊoQUR }=gb~ņ>eFKNU#Lʢrg]naogtqz[mܦbz(xvu~7Ga%* A". p"sžFn9=Դ&W %S,ۜ&JKcBU;WG9*/frVۆ؆75A4ZH+WN+ܽrhmgtpcS+P Ri(CV mַ3_X`/W[ݴy]]R\m蕒L{ Je0{SMjmGQ~AI'7+nھhkw3:P$BY8$*Cϼa9vK(p0S+q(73s:i1@×6E/L.9dN;TF5?tVj@YÆnz\!zfNu`v6} c^ZDoqu }̜roN;TqU>M\&=&_;!Ipa2s %9porͦŊ+Ӊ/)w>J~5?ndβIQlq i?ƥ|rA ]v25=__"6xd6r?&ubJC x [PZimRg c2n(|_1w>Ѝ­_SnRde>/^V*xŻMT nzxÝ]o~W#^;)wɊɻu3,@0BV7|~}̸ˇ[sdm|)ϲ<41aZAZ~Z10\+O:ZJNm՘o PAdo39k51 < d/  WqnCfP0%Nip68o6 uT y$*Y"wKr/}~襬H4u[(ZVgx;ђSl@cQ$rչϢ]]l,;&WZ,FٽS_lr5?/oG2Ůu-4^ֳ6rqs zh,ޜ_BMql,6l ]`=!m<ɕq͘2-bgy1?5ߍ ,U|@zsSƼџ¨7LjelU 3㨪M,{7&ؚ8]|2U,ڬQdI 04}mC+&:/i1<=D dK u2+_`T`H㨴ðR2S_@+Kòi()6IT%O9 4+O׾q%2oE78x2[x8xk7`~c;rFL J(1HkׁgQ]x\#9%K1S .SDcK)V'8q ac~-l;&Rzv3EClF{M)omvt) S>Z_Ҟې#|iO#·_W^O+^YOWL6Z.e"6cXP>8{B5 3>2 JP Br֗?ObݽwwImqGFRN86p]&dpk$Zsw\+%:∳O㴾sB[+tM{_0P@0}ҾIn"&:81 Q ?aQS> >t1c>PRn,;rTj2,R Ƹ%kMIHC@+!>y(z(]ùswR7y7Hnἑ&Y>bf1g>Rr(<' $/҈>?@o|@VQ|c?o_BW E1bPa0խC:v0,  ˮnQ-WJ}l u@}] Į|qm(|ajU$ ONr/ \W30~+/;Sii?Iy/%fz+bUo^}, 525e]Q&/ղuOen/~l`X9>*dS)<[DB߭FneMMߩF1=q'Hcy2˗46h2keq}7Xѵo ]U^e-z>DQV pqRp/'`oX1s~~臦:oIÄVRK&JhHI!GxF̰\ x+-pFW,eB};F5bsɀkiMeg=B#w緤T 2f[LCȥaVAVTFGaƌ=]+1@G 6}u]*RW߂Sdc9e>@S^D|B$էU!{SS'Ђozic%%)YbT aw P" 쪏6*c07D'a:/i8DUMAѵ8-WU?#/ş>єjEx?GxvRE4;&W"8>Buz4,&~O55hp,V|-ތ[] y68gf*:"(z5$'ߙ9h"|*W9,XqQx%2ac!EEt/R`4A|6_T`4E1g*|1؁{.lny͌PnPaM‰q1a9R1 S2*iAR)$^3.Ū=cg/)WewƵ/4$1x,`Ă{J&} b2,} |w cs8PgcMF)K#L_j!X D!Nr\YĿ ^HCJbAi$Ykd'0 p)e@a4QچD=F%r/'*8 0i'LVnZ+Z ZY{J~]fA»NV^jddC*>UlIJxQ~O] ͇G70-7Z430&k1N5 2@`gZxܸ_6{Xi~ |Nb =.`akD؞W4-MuK dQ(ɯ10 =Jx9*/:|з[fp:P,)y\"2 3hҌ[$X0#ɔ {ghܡ1W(P%/7,1w1FOksO&ꏥpě(%QpZsQ: 6!dlWyϋ.Փy^?hVym!5,` ĨJD0݇ nbjf>lt7/`HV+aY ϢQd\Z$l.B;о[Z'%۵tЈ%p\C>ލ@_~[W\׻Zo)C`hAb4# \_յKY6]_oa[Hj |^zw^YEcr/I (dKlK"P3Y w'_3j&ƫkUCѤdQ"OɧJ:bWëZ ]몓uU";u1jĊ9_3 9/^M JQh(l8)~ͻo_}ݷ盗޽;LԻWݛ^®'HQLn5@рUϫWMh[UC}bMmz)Rfk}]QڗC[ D_^N7ùw`ozk2W0 vYDުnJ![eEq@/_`.\1vHh0dܩPR^̀ d-Q2F,8$hA3B)2C2NqXy|' m;Y9jVp+e`HbHbDm&`_8gCk:l@] 'a?ayM5z||"Lc5ߎ3gWQN50BaFuVyXK)O'iyC"9'F_6 u٠<TV =*4Ο6mdqw|7ϰ;Amrw7/ĶW0o?bʩ+xzHgA5VP^1F^xf7p1l'cqpϕխTtPI[gB!1vy!`u*:Z+IK(PňncDy 碫n=W][mV钷P=o)2kDYg֡^}Ncq\<6 us%;}t'(1.%|xr XM̟y]!:0,L [,ΰVJ@LD@ƻz)zoSpޏmNu#$l[ g=U^.FgRJж3,;B',܇-zxNó'HWQOA.ˉ>nJ:W &9y hz^yׯjWv tm.h40Ul:\.B R@' ]"{"NJJx#baμð$#JNɝrN>U#"xЏb$Q뭨.Sx[ /ij?C.( ˞WՓʞӴT0*~ciDEKC)Cjԉ %"?"F3}ZcRMSCy8rRm$'uպhnvggs;v1}G9(ڇh!-dgQ'Lv)A[P2J&BD(%d"LP2J&BD(%d"L)J&BD(%d"LP2J&BD(%%d"LP2Jf1L 3*d[ou\c*jrTy]kO,"5K= ~R%B 5Rsl,eFi,X) o'u.ُ+ c>'Dё+*܌-N\nbNoz/EpB}*GD4cc?>wƏ~f&/t鷋8/f闵Z&Gs//_bӉӥKuv[-Po{V?6܎Ds ^l:MX&fQB,ľ5nHnɗPKnj? 3!Lyj:]0| u;${Ulq $TǴXSuQunbMQ71&Fh B>ۏѳuL(kVt4 $4n⮬U4*ѷtZ;ToCj.ejOWu>w&Lb3[ wĔĔObJLL)1_/RBHJL)1ĔSbBZ] -Vf"8{a 5WgZÞqP+9O={ƣS,E"vMatHJse,7Z!1HKbq09&7X$en7!oggnH;hN/R/ltc$0+9S,@<(h LbюJBeO[Odzwi"3cxwB0Y50&v;`b|~@2 d~H?0&_FpI9кo }gXX'QJ'd=܁IzؚK\miߓⱒšf ˩[I40*E!AUp)eUg`g@ B?`De;x 菏2N\k!'uS|><K&b H[S֔5%oM[:Z#Xs1`aйH8O'E}/$1?\,bb |>\s%)8z9ese{ϓ᯷o׽B}>>\(A` > # H+Iydy寿6)= ק\GìhyPjk1c;J Fa2ӍvX77s޿+fDӾeo }K"c=KӤ9>QaÇ7؇7Bn Բ_>ft, t#Hk#]Ō>!~7/d8bfF6tmcU\hH|Xqr4U\hhz>/͋FQSǥ'\bp>Ⱦ[< :)=C9gxLm~)+ 6GUNln\̨:a(cBڥOMYfvgo40JyBOyRjz\z1'!FP{1R8ij ቌ0y&2DFa"#Ld9FJiU7A Ý7OvO`*[wO÷{(y]0͜5Y1c2`؄hj4τ̈8^se.~CO >)ӪҶwov ި؂iꞂ7]ݛ^WJv퓽 {[EL8k6 ;~)Dac$0SDŐ$As —B[S'I4$IVݐD<%OISD<%OIğ].cUt`8t4WcEj q~R s^_|ӛ]ZZPz:Rt 3?]^|܀Pр;%FvĥBtV\T%f*zԥNSSvZkځ :en8RJjOʿ]@5%.탏jbmSJ1nld>-]ymӒR׼;xKI[[<_N>]ًBܪtrEr 8S] =ٌ%ݦTfKi]["Dg5aXUn +Xkz"; JQpu"g۝nWGJ G &D>8LjK%;dl ] z[ zV"Ij~b<Q~:Ef}2+ȷ>5OٯxDr&~T:M~MKۏ}?qẏ}&Ձd[5ɫjٰ?#]e+lH9 %f#%Ș,L;td$$2K"\JlDŽ G40)m3Fer/'\KLDiXm37aBW*ٛ|V}Hof->l vf7K\\]Ŏ$,W5òQw]eQh %ְ @a#̱L+(hUl0/3@%]vt; ^T8K&y\"Y^TqXŒ& {ghܡ1W(P%z%Sܓc)&#ca E1c ``B&X x&IϢM~DD W9,XqQxhL4@΄AYMDjw"e5Hw2 *eF,qaBȘwFcO:;%&ǘi; Ui;:i?LY`>{p W -Xs1asL? ܅Kq2o>yϋ.Փlt7/`HV+aY όG⸴H*_\&7;о˓YN޿Z'|۵t*4gXz7&_HR~mY^]__-7W!d j1lUU])j[wmkm/sw/ 0i;{vނ#1خ$m_mٖl٦_NkI|/v[?@5ͩx;_w!?]MU R2[{3]ֶ(a܉tqN(ɕ龮X ݔD{} Ut;y6YѬ}gjrylne0YduΕt:rYDF\IJC~Ye" \l]cQ_߫i4<>oG3ûoܟo_}w?z'٧|z+`W`=uCh!Xo~m5DMF 7: 9b@)KE3 ڎ^UOd?pI}^,xkZ@S='vV Q$&0*`>W^0!u~7.rM^6JiMl=PEzZl;L$"YDVCSA4B11bZC΋|?EYm|Nd)$`vTdLX] j sA!OD%2AHjKa{`gٷĦΆ`<1lHʇK'9J *|^I_-<Ĵ> MgԲ-<41`sȆJI"3!gnܗxq 4ňI )&k˔)qP . $ $|㸏k,!9XZ5vdXڿ֓ѫׯ/BOMvc/.Mb;㡞˖{{q"q 04-UmP쉮([ӊSqր ATLt,광Ԯ}?sua]\M W} ̎(V}Y{sKcmzo{oԴ1gĴT?i>vO:̗9i,j;zY4pm﴿u4-jp]dmf0(E'o>&FJiLwYӛ ON,?DZPF2  :pPwBq7n-FX1rTJE[o_!!*) K=' %ᆦN7{TFn6bw>HЎt!wn`9x;ڲv{%z)Yh nFr!_{=v@/2/J>piI^E˭UJҝe[~v$z:Q)}x1+-q`dvgj 0B6`w;VV !2GBW2'tv.+{={.ojx}OBX|Aep8EĮjWıVR$%.:_,r9_ȯO#GO[T\0ٔ> a+3ɯuKsh\2|z \Jٱc-`OVUa29޿3a'UDbU۵пEpeėD1.>YF8f/]GKNWո]D+#5u.0y-sUE˾Zl͈Ӹ0k[<\zJfd sn%F,XPX(fV&xy~ca馻Wmmі6w/`v:trALYUQj_э3{?}Q(9)mX2wQ=ܽEI&{P0=ճ̼=jk]y9 Z/LS}ym|Uef\So۵❔kғ/w$TŶ3to򊇱 tn. ~q,ťWͶA'@Ξ͍'\B-Ɗᶘ\Q < =JSÒYoU|_:Ǽ;&N1=:_p#ѿOw`JVr΢b1< &&MmY«.i@`rqo]± ۤ?~&KC0 kƥ$_hlXQ}h20PNJb wCaz`91Ttm&=jFΚ@@%ÊRLa1RgMT-&h| T-upa Ohe194~TA JE^:'>*= Fq&5{>[K N='@sb\ y.PE*]X-P9cJO -}Jbp+yz֤0kP8Xc|mkӴ&<4 ?^0w&*sN,<ŕѷwcp{egy׵ mE%ݝEbz }3^ ayCC&Yp*;o:!y7k&3uE{wڒ1ͽ-{7|{'X7Ma'__]1NDp(HR#!T" 80`4@ !*3VpC E8 yNđՉCwPw^nmE Y9W1-s`ȩrz3{u3FwJTPRBJ4X"-1P#V)K!WBP0H+AHvdw,x\WS;heqm.9; ߶@ͼ8} KQbbnS BhАJj`*,Biv@6ǎ?z^{Yؽg)׫3囧Z Sb KXZeLIe`8Qqj2  ks4i2taVlkݯzo3-Ψ~یai܉oz4m`Y{;VTωK&Cb?i,yq+8vZ a[Vp/obZ˗7^UrW=_2_l8l|Oބp;ʯ:tѹ'Ww7~~3wuru?-,ۍ-["\ZO%h8MVHLjr~5kFJcROmP&qxA:DL@.w> !F֚I9vqIBSěfO)Z{>W7r\ϐʚjj\5r+f̔$L۲lmM'bW?غ9vXG#˿z㭤QtVҷ.X'̐N!To`k?)`0f4j|5ez4‚Sa]ov)._Л|nh%A7;SIL=BXVK).#3/m"{0~ n4N9vTv4_x߽(^yg捙~g>|.܌&GuX!zt> uG?[ro+k^ni"@z.@a7R"zwe6/rf›ۗM'Zxl -_k n? &/}eD?x mysU/5)?/UnEiRhxj>N {ߖՍ2P{ sv4,sA/Ok{GL=e-z}<45Pa_7z sZD)2lj%) ,ӝmBtg#9r 8M(x%(A[R.#Jl(hF:HsUN:.eqQw~܃N#ܿ3D/}"aUK2* U971SotDf)E^[0Jl7ڼN@ 6mMIW^f1 }2} N۽8s zP O[1C* dɐQKd%!T01.1)E56{ (WKM).FҎY`mZb"g^#αi;X3ʰ)6g'} /TLbGu|<~tueZV`hOvB[J7!cY`* OA4`5& :|0Ϣs]@^aY(+MڞbOX,'X8PDךl` lBdN S2 <&?OoښZH#X@I V\0O +c8Q"S`Zi]H"I3X%0YfYFcOz{2J]ژw¿M{G!مzą!/Jpb/B-DX/d$&I10kܧR\ ӱq13M(XTOS{=Ħ1Yq&ckO,Y+pbp `!1]XzNÍ3P70i[  ɊKqo2˥CRe \h] 22JQKO#d䑺O6\\ξ~疟nwSx2Zz6LsoɪMCGLe>Is{5l!ymʔ*ʏɝ1o'1?_'o򋋺q3D%o0 Rm-P6|7k3\0w/B(eX7R7QFl'V2ua ч(Q*btIӟV[j5mTU4?:PH8| +KbD p2*Bkz %E- ߊqk~޼^]7^D]^ÛW|h^ۺ]#:F'MhS^E*viK@& t/whn ҏɻWCCbۋ8?ik.k&f_ tMԟ&l"Տ+ReMxx!-.qr|GkM-vTf r2y5ssi*QR0>`!Zgt~ϸw;gIiLC d.(<)xx4S`5pDwm4,(M\: ecqנˡ,7x2zCCrtoDpb' Yk+#mP,A91>lom~ oNx ,H1Oʇ1χSN\v?Midp y@0K 'CXf*dwQ].jv5Ef52;2r=ls 4>8Ijӳ&F<)nKghg#SxRf-\< 5V vE!#[>^g^Uì}+tp5ό RN$UN9,AS6Rm  W2=zல?*|?^B*.nGM֭NOMtFL11E y|FM]MM(|LƙVk0gDeKO ~I3s+jw>iKkc͝,/{m$h<MXt~G2Y j)+(Hl6G67NȺ`m:YR g﮵mn4v!-(Z'8n;C{>zjt5˾OFGM[h|jJ6T<_ '4j1[E˧|֚<]xٴWoYOpVUi֖;KpZD[[|\*l$)C--Nz(1۷*> 4 :KOzZ͵IXk'AE|EqC5<msw3zZͷʆO5wXZoIW6ܼUòl.n/^&IG64w|}hiTR{~F '=i_;rJ";Ϛn;s$4~Nc I_i'@,ޓڕl/'eM7F+)(tn7酁NgJ`HV4b VI6vp-Q-0*(EaX CbqF&M:SM&W$%/rMt~~6տJ)ChAl0x HVoGC?W\1WT.9Θʷ ?_'o򋋺q3D%o0 Rm-P6|7f` _2Q*nfJ7ȐddQq' KlJTAnu +ZWk$ >?&.6]'D3ɨ@Xc(ArGdⷢ >R|7ИL 6C <ڝx8uN2fZ;9_ASkJ_FșrЩVAG vՋkh19P`'e+7_>%z6^ho֟aƶ2=:&?|%fo3h>{Q7LAmgy|H׿ʭu܏F+ h>ނ\pa>,Ēu{+ɀ?)OJ s9LKĮnif!Z}s5Uwӳ·(mj 0Kg k L{ɱ6H 1̧Zr:`<3; oȺc`c5*CzQ$#/P *Pye\L86m4V@j!>I*ү^&3RJmkm#GE1/ś`2 `A0rGJr.VKlIݶ(+jV._ŢX3~Ck]G UVO=_XUe (et]WWg֝퀴N im ֫?5Lf 61T wD15:l1/ VF+*.9˲ː1%m2,Yg-똆,Tt4X02\j+gR g^yZwa&Nuʙ|َ6Kuo+dYae*[o[j5}{z K4JJ)hA rIh1*Xm nXtل(孢-z{ ;-7Mv_oґ?7J1Pgpb7›-voNpivy*/,AAί)\OctMiW i ! ֥H^gmm1[ܧ9ϛ/s5;c8i!">QD T Ub~rvhs0>. Sd[ADF*#41e RP7'^a=|p"VMjTL oߟeEYNƁ{ap 2Cʂ"j+\Er<9SsF5ejk~XjR hEy:Ҿ> +fl"q76k,O/uRkxLfo*au%$+3f&pcZs. jV}z}=ϻu=[|줁v&ɇTAvZ#3!6]"m5~^Wp ).hz=_4Eq3`S.R)jg` {D3 =z>dY m󘋐/IS-_rgmrGۊ WF],6OǡCM|Y/yA𑇐|{Hތ4+`e)>Wc?pe&ov] Ñк==jZnl2령x.`o:`P]*ex<~bj[uj_k1l)oKbbf =/Z N]K'j}=}$Yw<}h gfÊI?^b6垵qx?m}hwgII]P:v9By6$%+x`!G\Ha 0eZSjN7|9?ppvB8fJsMĠ옣dsWzscPM̂qԭL0Rl=.E38bJ8$ZXE^sɫM+ciqYRY.>&;> mqрW|zFd('#|tS=tPDm"56s !ea Ұ,va*:SeH)&O/84 MU)&toeбA^Y(U& ٬4QE|@RI1- ,-fPCvLZ6hڠiԴgaGP Ehc:g1C//5x[f?ڑīw/i<ƯӝgQ#9THZ~x3_5is0h&_[ trh$Khȍ&zvb8v+^<|$H2ɷhyx3ˌiLP\I(m^fǟ?_-B[S{n =\*OobPm3%BOY 5fR]CT1O.3!R O?_2B]_%OlFHEie.RE蠊U]J5BmI4%` PrpAb*\ JI 2@-[tYeb) ԾiUΑ~-g8?46x$1RBQP,"+P`ݏgV!3o]90,& 9x'%U\X| 5gZ̧$z5L[oMj0%cR#L҄R#&C"dT..ABSμ6J P9[Z bWt1M <F9ʡ'V4[l9,r?}(Ta}m} H tPJ!B0I`cj3hpZF 2s4Y!;t*[84D戶Z@IҠQj?}(T`Q^a /}9@M\tڜM;:A vubrx5)bxD@ٱOUҦPys:D8@jg,EwCP`ձ'Y<#(k6g*kQ h1Cs5I28;AEfaL´&%dxiI"Eu=| 5VD@& (PyymW%#z$G/mUBt"*V6g*̛@)!%|7a Ix:߇B T ^)5##*5蟉jWWׇBKU7x`3QeE\]wS¼Ԧ:)ZJ¢pe'/1,  bj0YQ(a׳!VI*u| UP[%4!.Qx`Lf9 }(`^0S{S+-3p#8x40h,͡Py%jyHBХFKQ| $2| 5G\[7&kK ٕW("&PEC vu2'( Bz C)ńb}(`^r]yUTt@bF#Y4W%RY`VI1}(TaTGx y8Z7zOCK=.{Pf$!iƘf(LJP4Pr=KV25i,K:kWś4Yl;&qOuXwO4]zq{L8T|(* F>W?IY%o~vH?{fㅦ^(?'7䚝?xo[et3~wE #>˫e3BY8`荆Tzh4jf'qe@L ʩ,Q4HhA٤|܆CQ;|үtgCpD1Woꓱ5l:d#8V=)XWYo?8gUC NhiM BVPEp%mPsf,eJ G iip}slT d4vm|EOh4~᳿;mqru>kJؽ=BW!`[1?b}3;l ,ʻW"!*϶ vf7Tګ0&]% Eb $‚PЩ#  8W0luI h|rL9ݓU |7sJkRTch)bD8Ńc%c '.Qu$L3(ÚJC K}(`-&$gb e ^\| 5L6xJB$`C -+*َ>0oZ-vE K<ɉ9WRi]>*0#AJ4 QK<Ge(4Fu렃0;|8nJNUv@0e>j0Pj3/%4 -ÓXJgA\ 19#'6踱9;(P`Փ{ $B;i&EJẅ́}(Ta^qS4%S'%)UJ;CՓ}]o#7W46`1H`g 32AS,K$3}KUlS(z.fՏ:j48w|}۬ ļٰ*NzCDrY#!Kldhښ"qI{Sy(^2 L<.R?ypBϚoڡ|-],~Jv8>vCN);Rꉳk(/#uC >m~? V;߃Rwɱh- .IN3)(3$ 5pE"+^@cIz\¢$)Jm|ksydSDXF-^ qo"j~"uQm58,VfCNI˨0y>hlJC#x"z""_OsžE(,f +܀2ƁIĖ*f T XuK 3jBjY:+IJ*y]0EŤV H5JfQ&J\P (0l*KIGKʣ\P_O (@ōO"Sr%H: \B|>$fSg8iBiB?>gXYwX󩏀Jnogq"zOh8fqZKumM@M4mid3l7Td v奇 nd`'YnvPjEs0eҮ֦*ߞfEu|Y Q<1C*&H!&(ŨVbT KKr4KyFzZo1&ww{i 8ߴpsw;reւ]w({3ѧ}o.=onwǙ[8c̎1Oǘuǡ;Č13`'1 鶻Wm'o-4llB n[Be=n񯝔;')qIo8^lSg(`EGSyj_?8޺pgpOGR߄XV-G,:6=fGL, y{/-s 4]&.!6Nt;$iA8TL (YHC^HB{@Q"&t<;1 RqQpFL8M\p^($E=Y9Guuq199LQBEyAi}β"˾q[?%i㩌Mh:~\r=GTvAs(z8%8ZS+܊пPɎesNKq*"AKcm$ZP 񘻠 6Ks1AbX'-u=DUs; ꩮ٢7O0U}7 F)0_fVͷ[f 0 @w7gƨ7vy[|rU Wa+9a4SFhg`2o:$;QڔPڂV0EP[ PJ;Hu"H' gs0Q,O7 #ƘC esGߺ\XU~^PA:е@f! -1#hHk\o>M[1 @n#F:y"I hG苺Ў,>=?+>"rZ#α@.jk#O|<6bâXxLG A_2\)#^xqʤ'ie)p(ڵ! ZO߽#wNmqhXr^%,ʕu>DxmAvGo;71*ҕ.[^RHrHD?<ڝ緰d? bƓ/i sN0.o?FqGn|9J@hSz.zUp8U`9M BFC-E[ipoMs5E^,7 [VPɆZ-Ro+Ϲ޼0͋jTq0M#Aۍxz}j}?LC͖|˓^ ÿ_ꍠS`FSKu,'0neR'R9I4;AN89:uԽ+/uet 6J3A]UTW ֐Ȝ L3B ]OJLiJ,qф8u! 8"Z`(AP\?=+QWIb'UYې?Q2hՂHOf$`<=㒅MIa4FB;? O|Go5b(002&&!'ZH4 b7DuI%y:F*HNfC(I6N{c? ӛ#=0Q UWX>YoXE'$IO>sF?g0?><=riTgO /Tۆ;ί e2Msx"|8)Ы{|W1N~@]šA-q(+yGs H'Jǭ@\x1dTG<|B䯋Elowo+AR_?^4#) WU+xdeS䋩MWbғߖP~ͩnݛ7njgu\Ս}~+G&Q_c0)jäpz~oxO?o_?ӻ|Po=H L6{6<3XY߮}iʺZ֛-4ei宛a]Fr˺?Yw*dkE@Cs9/ףۻ쭢?94`ҕ0+6m}?f /A17smXb?E/cƵėw16HiG.NJ: z3h{-^ͼUG 8CQ#cAkHP:aPE>i`M D юX%*jsxmh tؙ3ѷ5#Άp⑳jX ֯Dlk ,~7n;X^>/yS.!N C8YFs3V:N3=>ء~e77af2w)0ioy0v|2ؔ7bG QF=LB&/Ա %gsH<&ha-e2.j9Vg+`鱶6aӱ8;rz3p# qc궫"s =nNǬ<n>vU!xVHhg/Usʄ2bV =,UIXw ig ꀱVG#$KS%JQ 3 D<Spglz ',],}wvYoSԭf;#`ܲ}.tVYuӁ zf9D !R"ևl*v3p,i8nPW~͈}L^4dndW4aP3O߇כ30_2ݖC>nGoWdu>p}[gw-}puٸ݇~e7H["^+k^J nCVK%^VY Q<1C*&H!&(ŨRbT KJQg4YN]M@ s8ݮtgH Fl#Qn,aSÏ-~D9|\貝.+P,YtdY mV!#J+_1MTX0ҫ E6$RVn;00Hv2M7Vh2psrK.m 82K|Lmh]R3ÉN ob|}D8Y}v6lts3@}EjFmG[&1a7mf)1M"ǁa8nx2t*- DIѾվԿ,xvɥ- Ӣ&%p8qҁo+x)3ާ'3?h5_J4+0>XQ1WҐ"Q'o3H(DX &g F8%o<7oƂ~*E<+Cs"_Q$wAo!ZꉏJ+ӚDpJaV1vntxR{Æ7;&Lvʭ!je6!HK}\>λ IR9!1B^ qK)qbD2K6vBk/E8zBC>W%i/tZdѧA2-<L¬>@1s2K$hXDc^.xAX@/’zᠦ0%TYϢb ;T4XغD4ۨV\, 3`,5 ΈIR) keDurŬ"Sy4,V(P!))D%!NrXV"\gTҕ%DHmK }{3#bEӄi'`u9Mw.` C (ךZ V@@΁JvhÞK -qn s,%k#qւyOmXIo$ց}2 Zݓtֵ{/_A٥xl͋ҴyKgܽoh)z v9кV5WbH`zA 03 |) Pj\"~_aďq"~ t4Z ڞGBL0"rZ#α@.jk#X.!FlX 'Qa̙4p >xud9+^jv58X١Cz!^N-r<+[2Wj5բiçtp-F Zc1NdiP#`ǙrTT5WϑυA7ჼ@K)\SE 'A\:uђ^ԥUĺ2l(FGτOI%?[C"s&S"'47HъPtsUf^-vU]&=Zaif~=?9ͱ-jx;HH3`cDxԿ-( xw_w/ʣ7r_D+u^]-|X(~<jN80P74i^a6n *8 헡ht.|W]O^5 w3s/G9n)}d2Qm?~1삂Z03Jq)A PUZ_aPCA㞥 +~;QN~.SՒ uqzUHWbŋt^u瑘.1UQb8T)pr5%ψ2$Q92#H@0Xi`N ΓSR{O$~*՜9و-S7Rx+?=r#h:tg+fyz>e/S{F#]?Fq'[&}9d<ͧYM#u }Mǻ2|kFnƇ7{E;.8SGLe`h|p]g6Ԝm6t7f1Ltss"?6}eiZ.dѡ?n{S];{BP#Zاtii \1|I 0WJr) >Bj*=B0Eq<Џjy<[s:7* FjAoȫ7 ʿ\#᭽2$JCD(8 |>$fZۙՔOTC^pT+^ǧ~ǥѹq:gK 2݀4,\T&=MQu?5G{yW5z9{s+d7g(Y}֜iƷW,>k+kf98_a +>&QI5a>֬"gQ*RgTF߫,_9dό8'U> rB@_0;-oFgQUCWOM03C*"ANhXX+R0vXk1VjbLwnǻؽt;|];+9[=BzީUk' 5 Z'Tv{=L}zϲe{7Ce }0h4,xx?o'n_ @cC񾃔HDf[쐳"৓;nJJ˸XD{\5ʃ>,(wFh~{u;78\lH*?rk?;Q}"R&BPS )C8 S)b͵?ܹܷz5coꨩwYV-`H9p`67MHH)οC'JZpV>cۆK* z93\J@+<\Rtk)!+})bOnFGnöx'y1lYrH_a˜Xq?Ċpa!nTsR32z! + 3EβjQnEΉWg K`RZ-}Jy_f ^fe:a.M;Ѕڋ=vwN)랱P<@ sKa:b͗PLr ֚zuv8y&4`m 80 PΜ6;v(C*wmmyY`G>x&@rp"/1J)RPS=3$7 o,q5U_]KȳFΖg|~FWɎ%}:?W&z5ң];G ˲XY&exjP !2ϬV D)pVt+d ^u s6-cKS hsɭ,")tTC ;Ib張}N0ƅ).e~+eIG?_3tY`<8q>͓ř]a@"WV J1uW&ý|' _Ϊg@0d4!l**#,IVˀЌ\&n*'M0,o-ai6;?L2К0<>kI5K0U/aGFMF B'xWg+!|{W-Ն G3f1tdi8*AGdӨ練Q2>]1+R(5½QlN ?;~7ۛ_n0Q7Ïof`R?H$ele mkho:4UlEo]MR+7ӊq{_l-O_KW?("l)EKOMЊ tz{~ ~. DBU>*w˻!MvCC`:ZiM>/`+J*h;AN3JE^QgAA&Q j QRNQ:aeyV#$HmnON@NXOceYZ%ІR& B2ҀFe8 )M:N+uK ?I= )wh j TJc%O+"%){14g570̗D+frPp~~Ͽ6g (6|2,wUw?'# (lqxP1jŝ1BJ`vp8wT' ŪȨͮ;3~ 鬂KwkRz42~>O[Nq%DPsi,Q{ƀJYA`GL,[΂r}6Dh6SHC3KFDoAsyIޜZ=ۋ24YҔ8nB^ScP'Fh2Lԑso#7pYY=,YydcP3E4^JJx;,s$#-[#gn0Pg!VyZls^X/Ycg?y[`:1Z䠔P (H%Z{M@D{* % {V¿WMHpM6)tRC-=v~JXWNk+>tIaj'_HSՑصͩV$k\VX_B Z鋱Gjg@-go*=Gjb#0.TءA-X%T 3ɘ:3D8=YX/鬎9$?QΟ,}O5tτIɝJidT"! KAQ}[[,lӁ:[ء_u'ԯ)g=8Bu>غ4 =IMbBob6׫*ݎ-OfD13ؕ~*3)+T';;~Q@F C ::hwZd1 sx&qUI rH# !>Kjʈ)15,Cʘt5r9O8?*",dzϫfh{38+DpPL@;!=i`W%!mG"NIcnDӘGѰO`` 4wFm6*GT KIPL Utȴ,YWW Il_0{vmklؠS,ip%6]*ہZPTfĞȕ' FƧ DFgVHgxup>Dp41lj($ASR =T%hbάDK^Jw<2L;%h V2nY)$SNc!hCLcZ#ޗ[*6gWO9h!u>7Tn-/d!\e;NtMof_m,`WVܲGt&[ xS:,`IIvtw, "4 *Cw{+D Z/@<ͪꤓce;\!`vy:\Ҭ; m V'G W"//L ,ƱMDt v30Dtڶ%=dyJ&6vsLNJnxDrҖ'W&j=Œ)Ɂ LBu|k(ZͽCa6{)ڙci ZmfYDΠNxSHz^ʋ.80_ 4=.vvRNDN:/I.ڮ<\ (p8mp҂m -{BEo-gM *LVXw N&Pًwz.Ht!]>!6Uưq2q2?g垅)vh>WA.Vo|jX(zYR/ƷS~99] )à(ReZ30쵌FMFSZDٿh>svu> I XF|0arroDʵǍeyèĜA[N!XIa|+(z*t| 6/%U9o϶*Gk$^H1 N9ťڦ^Ȍ ^`A ƜR؊á3G1ekl0= m@ og=hd)G JҾ@tY$h) Ӂi]Iv@wʾ" \HL3 r>0Gd1mH3K'c" #$$8XG!1z ",iP%LR#1s:ֆ5-;I$CJEKbCe5ZlQ0yC4)dyOThtQ|a.F*g.Y>zq.`9M6s"gZr9Afx_o ;=K^ C_@|]L)^[c@!fH @RZ{1[,} *3*=OQOY]~"q{u߆;M٨ ]W]wjNZZ'$ms*2J4}j[a]Z|=ֻ]7z7nǴZ07}A[VNij|Cހ=?5?m o[ġ}/:JJBsL1%RR/ʪ>J{Kuw+@]xθp!uCjDfKi4[ђf{5kl(bb1sQ]D`н%;rI ;ѪP0T+6^ ʽLc8Rcib ")1OOFzTo"T։KD9}6_=ͺflY'+ 8Ɣt#Lz<>{q$Eѧ阦sحf0]{ <]ڒ%z77AR2%"-Cd"ʖIHF")Zr;iU 7Yi (WKMIL®1jveZ2h8zbqăt:*1R+XlPAVI@"(4 R@ل  mP,g* p* XTHExldƀP D4VN:/(1ԝwNc&WZ`HE`6@@y eLKbS&@cgDcRs6kwH֦m%g<;PP.X+ o9Qix`,(U0xa23Ń N^J>gmat?~E(,|MS`-@ZFlUlfWu}<Xip([cl/ \k?nFO_VWߖ -4n6I"fJk%-&Η.ȧ}Y':::]XɍGsT̂Ll2sĪInӢ Jтd k,)PV8dSo#6]nF3D&sL(QmGs pEw9lVJHH!dHdD&ۈvKZwY=ν +6wG{"n?Nsc K?. IXbcs;Jʩ&PS0ua7,r0hDñ3O @;k,xDdO-38uRr')ÔFF%01j>zŭYOdY~#GI+s(-Bv+h]>U>D*d :Msr1Ǻ \/X7AX2fN7xNJb|#ûM <.@kYX\`-t{gn[D6"eCSa(PE;  sx'qyF,B0#Bƒ2""&ZH0<)c"ҙl6ƥo!W04o5O]N;>x%wfıZ4xhzMc|~ew嘡!>y1wBzdӀK.-!mG"iټ9<4*k4T@R `1( F刊`) 9U {3Om:8,U٩CŽ_/;`>rQ!EA9x*1FK0vp}2<,Bt 7E&>]d"jR(BD;统sWTgjߨwb TJA*V ^X5KE,Kywd"H,o_|fm>w3[~VN$+"ϾM5ԮrQ贽a.]` (#5q|8]%S7OK$tun_n~ڼZ3m$Z~̍=J7)&ۤlE4aRCX|ُ"XqA`E>]Ɣ[|CˡiCV3ۖ u*QȪ@/2v]=nBRyRr0=w_ӾUin9KE*L<ɡlJI RDpCJ,s$#9& @ s^;asJ҂IQ*`^gRo-~+Rؿ~.ԿAay:ܲ m@L{ 6c(qyS0t_C$QԳDjnylid&i%=^;'.8.MI8Ēe m#t9k<_Q-N{{fEGm7M$7v<]]nYU7zSuutm|?{L$ MԢ"ljy3J]<wJ aȡ סWr+ǷO5´pp<8}ty9x.6.9dgM'm=SCӤhₓFMlNu0 a6*'.wBqV#!]APpU<MB>Bne1`afx?7p{T6_WE*ʱP}Ba/x1^Ӆːr@ B*UP [X1cP豉hj4BZ"E]VksN]C2$ ר]QQq!,$S{+%R.˶l܆mut~]v =6b ->/%;+ %Z2|mGmSbrbdF/cL*BkoE8 Akʧz^G*n W"ؽUL&b9I}4 j-$* %^Fґ0хT9xA8Aʾ! \H☁ 3Pe  D9#ȁf19(쉋ei8XG!1z ",iP΁:KFbR4(mҎIiMaG *jdQVE<@rWExKFο 9Rq1U]JT^%ĆqI*豂QRҲ<($3e9D.!$`k U cJ#68ROjRz\Tknmr)ԬY7q</>箃wa(I\J^FYO}/.i߁$RaK>ݮ ۶|~ξOߨ`׮CM<{K|󉣩Wɨ\#TÄ6!FKgU<(fC^E4:ҕOooF-Z)w5e;@b Y+JF-8fݐ돽*w"[Nנ]evesgR67ˀI4Lt B]CK9cHrrg JA3x__npC\|Z|O6 3 jfP SOAlC]>d-7н`ϵ0{iaB4WqDt:(6GYLX:rm(-;\S2)iAe[VѠlh| VYH1Y{:wg _~y'#\ 9E2hp4FDQk0D{* % {| ,[ovAgSV0j#lz53b71;s"wfH> .Ǔfha~06_ud#l2FTd^kx26 b6 wӣ:HP2㺸_Oi #Y;W0EfX -.?e2jxetU)ި/#c"/_%l2h|qwR"` .,6+Iʠ^LaKʛ"qKĎQ|c NNiBNLd$!#$]Rj?Z7W]OZ\_hU.'.%ӺOtm}9 zn܏lOŕjmgyGބt)|iY'cjjJP $߳ˏ { >zD칉Jmpߟ}/Agͷeu-ϋ*Q+@]x.Šhݏiyޖ-ggي3lE&{9ئb1sQ]D`%`;rI ;v(joFI*N/^[A)ıH1WOpfc;8=?O.Rd)>M yd`6:j(W滇أW6aV*SfRDzJ ֔B-F9˝*u+¥&$S&aJ;f2Rʃp-A^jFeltL=UAJ9J8ŵ`+*ḌԊ1`UJ#cmE,P6aƒbB4Th=7|#2\*밊a,%(: =gz_ewm,!gM$NW[Yr$yfغXպX-K [M5ug}EQNgUUޱ=-So\,MPN}/T")yIHTR(ke 0\zz떵7 }[[ *#5H\SWgr>pu<:SiWApPӻ\q>W6:2eEfژr]QîU_~eܤCfnhK0w`NSACltQw} ceqA?m$3W{F5+\EBVHY|>$fӻ΂RrcI⍔{m'IϸKLs!76O⭼E>/|&Kg8_RkL޸Yl0fu&A\ŵYď^~ mme>L: W8+Lh! "Ő8)%΀Ӹ${-Ԓ"?`83Pib%щ KRH^ YhP$BHgq5p"p N@-I4\E_Zbp8z|\zG#W*P , G\->E%#\g-9<hwan!OwHгFɕ@@F()5FL}=E>4Jr \BN u7'*mңiǹN~^-2Si>]$B luoǙГa{}qE @kVqt+gT$O'=qnNސ|!b{C*1()k&!dpF oԖ1I !v;/0!(N+gƐc.MT=@Fș[@l6p_x grGZms,GEϨ!p6}\ RЗ?=SI){}&Wô#@m ϖD"TT{JS+FqKio> O"~,2w/jM4auL[HS69 BQ)(&9I\(ʅ Z:W|^/[bnpuZVtw',5^[$2 $܄xh@ Ig=h7|A.QMb2xP2R"!+25H6\xUF- BLd P\B  nWyЇ]-[2Q2@#p1"RRF@,!bW4QFH8$(? 6 '(`@Ƥ|4*5T h Qh4SGT2i 07RT 4K:}+^)95TabUu>*bUUW3.3  G; T (8 Թ0&5MSOyPѭǫ\fWTW}PMO3Z"-`2G9/` P.nŕQ/ho\b4F%xs{^%#gtj=!etByt\I.G"sr=.Hb>f.`R|u~XJ1u}֊dJZ"}䤎~am58JMG4ݙ ϻ"gO/:g+%~S{u=Y[fv!Qn\Ub[NHL 3qziy,2ZG U̖|z0g;^pF7 dlY5l>(oDƙ?xv'z o DDfrwo/>~??|?~whu+0>-6@$s?XR'RڹDH'i} X k1{a ԰볳4_N3},lE?Qy9g[AE_lw(>Q͵tInu>08>w[gL[W7}bW=Ǝ vU`tY"7')7TĤH1YjR ),8< YN6&" ߹\}7#vMvVX, j:nx*\\wYs). WG"/,kayaj`mWf'MDdRvn380DdmqY/TE˝[JMJnkYqDRVI8ބ鿫IpEG-5Mwaц{>s'1*LtInY;R 71a;1j&lG-y5t;N҂Dtf64u.mK̼Qq88i!7ۇ +O:X(ȃ)&"$|>5\:Glhs0sdr+L Sj#oz~*f: Ԛfx{e^U嫘H@o3H}`T,jY#: [+2^gh߅ 6m:dIA>቉$BH'Dvt\8n P8 qm5܎("*Q(Q$wAo!Zqr 5҉"ڷnFfW]&w..z'z#"&n[<9wg+oqSSX!!ZmNHR9ECbP+D*$rSV dRmB۰Ŕڡc>˝uSqF#?mJz'* q~͚%fg 9`K$ LP $&7|L+݆X ߋ(VIXRYﴠl4YT(L!/t>b>X9%}mCž$Ċ{E2f4%D*jQ,() keTGwbVQyކ5I+vHI;k ;Pl(Q]*jcTu ꓮl^Lگns/)B3jÚ >t۰frZxVsmX3@J7؆o ̆E 4=}zFKιxxkT%wůg$TF ^yh[QJVpmVi皚 /~#glzβAΡ(Rg-ӚRO  411߼3&H e0Q"X1_6xԪ=y:v, ,P;AOCw8.{_PXo_t:\9M~n|7gj1ü өzA;TP9grBftr6DG98qKry`G1'2T1WqkdelHU[õl$8C]i"uţec`+I (-Q{9[ Q? xЃrza&0kLWQkh޵6"SIź 0LH9'8X֎,H\b%[bualOwĮ"Y$e& \1۬yK.II;$A,nj0_~õ6rδUNQ,RE2yVkã&D섷[oAvgb6uoˏwM+_6:7{mǿ&$3%*+s6D\%Pk٩Ⱥo13]Y QreԤ~$ ˞)7q)t,ZڏE)d׎knJ>$B& X䙑np)%el՝ёy)x*&ϯ-;l]N=fqe송CWtGZҚ;j߶u'2_%]bT<]5ORc.+gQ3s<}4mYz~Fk[܂c[V!s>#{0@:Ko] XsG~_i?_ƞ{rjEjsBG %*nxzU/V\xӮ{ِZ"yW9d&VlʼnC% <9VgZlf:/sːu6dYN'[KOxF5D%8()Almt4hU.*ɹ^Z"S2!4PD.E[v[#g ʹ`G~\o^~.iEh*\VXO;ܺz< Ir8&rҡ1BGQj) Ϲ3cD|Ӫ05w6M"H-ɢ76H3Id -Ѳ([20$t$cɓV46sFDs lrML @8^% B&}҅jBWYFd*$P.p$LOާs)2FQNU07.c&)TfHGtLd ZZ-%yddlBI;Uuڛ d[  J&UN-T /m26T6~@;LR:c*)V>Wy%QL2yҩN%f!bHb`fLd>hm0{vAEꋽ͖CJ06և紥Ƽs4ôw8$~·+e劋cg:]F(;Bԧ^wQD] )2st4H _nH\#W=_A.uقS9JJ<;֒F]hU`6/~o͹sRӅ`<BR碦 ڜ.Tթ7UYz!׏ah4w&G;7gӏ$]ɏ:,\ue(?z`5knmd&"67Vss4zaztBH7L-铔)Fާ/O.\)31Tݸ'}ZI𻋥[ }v~Y{: M@՚%+g 1Ϳ\_0y,0//+@9jȒ>& A++URY[9&rȭuB'\'W/O_r],-=  (U@Zc 9@Nvc.1GVS@RdїD%Wu?A=Ƙ 46{klN& ]Я$IVŀ?tyx/+Gwj{S[;^IM[89ay1Y),TAF^V&\)k$DBיpr|3 ҩr@奨݇lu9``uT)p{gh8YH!*%),S"$&Gߥ-H2V 'Er7sٖٲ]Ɠz$i'Kk%]M'7')7{D0;y( *5,í.EOlҤj-!䲶;|B2;x$t\]Y:8P2QiKo{픭tQR6~3nۆ+2=tcʹr;MvBY >ki*fbąW{&a},qo}[M 6-4p$͛/묘Z뵾a0z&n'KVʴa3(آ !;JKu! 뤳^j}gngt.rfVԴ;q؟8i!6ra酅.l[ECJ2złU`\8H" T3Mߛ929%&i݌wL_?ogOop|TR\C2!9+ ن1wIZ'i=ņVQPtWsx`n>s%?@<.\{6]rZũw-T .FH.]rϿ/0g]0kw$Wwt糛9 wy[0 a&HS0ʈǐSΊOaޕ)Fx&DFcAɨK?TC4:ݦP3 P90B噉!hD=BEtQ<҂㚥O$yXHVY/5Zkb 8먕(JLW &*%&D#"9G 8d2q_"xI,t@a) 0Ĩ9Ȥ+U<0H+9lW"'vVq)T^&PĿ4n'ΤYԳJ~[xx$jR\R(x*]Bb)T aJOj,3 W.3gE2ܖl)=.zV2;m5svx ,-e/".tQTSW*$R&A6HK%- (`pB' Zh&xYzvtKisGo ɴE!;qIIe61xOHڨ&dFE똄?;*^~%Tm*-mat3ut`Tz {3zhS|׻iZg6Z<2},S)pyq|im#[Z: hjE@mɽVߗ 9IEryfhEhU+,yJ. |ɳ^$#~|ͪ߫:+k:~ &gft-ou$K -,B۶-œW~-[)ەaT29<W)΄b'cKn0/0v2V uƢq5,Zp;˳2O.ی}FmYCBl tXX|wmHW}݇ICoMv3/S.%$ץVʒUT2 a!5v Kn:-:eԲˋ%]uy(#Z !DCHBV`:Ғ4ؤ}y;qz|wgOSf"]xf Ůћ %f;w|Qj $U9_q`c (( @ę{ ':ZWS=$W#AsF萾INC^dMx"k 6C` Nl^i #sI2Άgn=wuwͥ[MkvOU67N'Wh)cWWF]d?gv۪ɝw3+WnohgEmwh;%[Zs͟qn4|5Sgb\Ӝ?7_mn =m|{0<(|Zq7W]3O͵&+mq [W=ЬParwkZ G"RsC2[v8;f+ψʎY68`QD(I13F{+{  4X)mO SŜi{㸡>q!FXE,hac`cݙ9;dfH*щE| .ȈY$*)nkeXރS+}crlC[yodPr@%^%e|%txco8'rQI.Qee]~QVY#6QH"OQAcRFP:uʤ$Y݊?x5yC- #G^:;Qxd X6f$+>LzCNR#}Y]Ÿ 1bS g,bϽ0*uڶzSP@äԇ &SeЯ /4s`¡R.be&PZXh.P፧:c 5.pEphiF0weS ӻ9WOo9ahM57dzDŽzIvtjzr4HHS\UCWn7O,j@䜊 7W@qKi3Ϣ& lrB`3@2j@GB9_{b $' Eͬ3bs%CVw$:N"G-e}K| `a-O cpb;5K2^C{dsy0ψio^SGIzfs` {s0Ϳx:2?S=աڼ_gd8{dpj$yr0JpaWVI N& ^M ]Łf?$rlYG'$QN(Oظ+of*ϛ܊pͧT_?G6KowodVKX}f9=;T'ddp=LNWaozBFţ8V|!N 5Ru}ofW4#YO9s}LJفl1Le=/ao?4u fY+EԆzցpQ8'$E[KbsIwmŰbYm,C> k0bGyEOemgtkl/um>BѺIYĹ6>aWkzE{&Ûd-7&~kn6{OoOݏ77?2}޼Gu~֊ӫ6(ڄ ?o{ *˛ RgC@|rP m/ځa{$};:NQ wg#4fi uC  q-z8V& qDIG7#zgīh" UDu%#1#@К;-R!JǞqyU>A {}a`'/]k T;!͂jDb#i+D0^V;v&qpۺBjɆdya5X$bk;FzC/¯㪲4kxKmg 6kv˶ȶG6rY@ET@ET@ET@ET@<<tFW`(9W>ر!LS+voHN\]@]Mo?+nRPh=CsyaM7fެ ?B H9,r%3V)H ;5Dk輯{~wu 7:څ$aX%yymAoz͠x Vkн0ˬs_. G~xu/*9tAJb*)Jb*)c](2(2(2(2(2(2(2(#e|Qe|Qe|Qe|Qe|Qe|QTE_E_E_0I i e<(㳹bZN](Ce #>#DClh`+opf< {e:J!gRy )~򳃟8 2I -vO:ODcS FEte4SMmcƙ %9S\$!!Z,KL5Y΢YD[~vfΎ2|=')^>q侾?͌xVxnjM5v;*0\/9&N(e!OQ ynPBFli+2&' Pv`?0Yhښ"uI{ST)bT۱wfg}~<M8w}G=pQnYYr Bu,GJ,fР~G9m<qfb̹ +. |>$Hu/MY{ʁIQ#IW@]7Gy !yD4Nk9MEYfiA|u:``̤Sps=[TPl{^[5|h uj+W{.!5+IkVup(iZzӚ(e_55]jftsrD6?*լϴ:œ.2y͘6. +ŝŠ˿GĊ 7"ni$\+//JO89| ;P mӕE"zMv8d* lIqpR!8\d3˜; Y?hkZɅ_^EMIb2-x2(UV)id:h  pN-8\QdQe+wo <"#֒T 0;vzœ-V{QUQ2_u3<9txvqu>.c%X}?'g?{r9{%bS>P> /'FGȬ`HF "QN9&C`Vy:lʒ b=mʤB}l &͵+Tל3psRNxagT^h;慶w5,\/5/k';nWئ{x&_F؝$2AXT3!) GD AC2%I!*9bf.t;ƖVz>-D]PXV@He SB@a&)H`4(4wAa̗=4 ј1mDkT p \)-F͍D# C!B4BzJ{NJ;h {X,jVu {\55 r{E{S a曒dLKx]}J0tt̺P! W[[0U XFVB!'E8xyg}ȜD>E90)STѢ!xڲ Vxkd;gDw7RvJ% J{i—CFV<WX[{"}9˝~ku[%2#eZܳIGtLL'# G_y^TVh"(WϺ s*# 9J BJLKIo _Xk*tJ`4(g1JI3dAMwep Md,8ƵY#s[͗b7`>_.-?^Sd4MYWپ]E ov#iAl=sf fp&x%˕#CDCegJ<s୬y{#V, ;15,ABxfDI9!4(=2S#18ʊO?/5Y?ݺbl5]]w|JК.Ah}ZvzD`wE[]^alجռW-`.vjy|Ӹ%ZƇdEk=Dsw,5^F/Ԛ[H7Vy@>u`| OAk8O]RBS }48ni.Qx=׬!f1S|?} d4Mta?j;&I{$і)7yQGݩX]jhy[k"Jc*0:UNX^S&z9 bs wHGHݑr{4 R!fbfeCp@P,hF@RK*Nmes;^diW.MO2 _RnFOEӤ\bzyuG3NVr}gG$pUNW;$K. {M[ g~t9_ U+'49IKФ`&'#2(A)fyB]051/:mգQ!DEaCք0BhnfH% "$ bu& ˷Yawc]h$[`0u= I֔@PKU(D*p&$T_LF#N}Vqd9\N]Ȟ!s4-%OxFWG1#}7i5CY9}d C2 \tE 2ڄ v9\^_'0 +Lc|pEd@pqs#;CM ma rg'dC2L Gǫӳ[w R7'9~{99Oǧ2èC[H F9$즟i|Z5䞧܊(=6K'/۝8'(^nPRPe0c*Q,PsKg(pB"lL[Q"6$o寘<$oc$WךP9f~Mypr'żV''Fx+Pj׫E~|Uʞ:*_؎Em efVIGF8x`J9B!w[LՑQc޳-'c^DY: (]AX͜ʖD)W SBSy,4m,&av!\R irXD$TH A%\ {||'1k8cJK %p8>TR\C2!R\Rl4Ok<휞vvx, =!o'\Wg f4ght+a@H&Ϧ}Bc1~S:RPA7|quݝ!Qxy`GcM;'BRWFLdFE똄 V79_2oiaQ>ݍʀ j|7veZ|?_ww<$׾Wx7?{|(!כ 1et.ڡ0E^FjL R] R"uHY8 G}nGzªPVCY{Qn@ &7'ۆ*]Jde7SX++]+j.`KK:;_`1|{hk| 9fv) \8J[bCy5/j^tְh9rt,n^Qy6#Fqu =+ε̠BR"eEaZ@-B'\fy2";]]9Y 4&N%~`! V4aԂ_y :8/>ϛ?.M`2oL3mh"K e g@l-_(a#h Vqe|`ng=MHREl7Pj;-\i90sxa[DGE>i(` i@Њ%pTvE'3r+fttut`pl*ATep ^ GROcdI0Qi' ;s4} o.EFH,`{n1T#toϵ1b5scDB,veIK{О` vjn1+BLdAL[L37@S2zUiupCɺcV0 &l bU)`Zb)PNLQLfV9Pp]'O'ߞ'FvK {Bb~m /ngm2 ,avE Ǡtd>HKPJmMN!S ԑ5rVw0X,u̒$gOgq S\ef~~g}?<=r_ s7__;z{ߏ) qE,wDY%=`8.XsUGZ *\c⒒%]m\ғ^*Rž~Y SJ{_C!,G::˔B]\n^_zŝ_+uaԍ|!rc(`i2{wcݴZ9/zeE>C^O_͛:[uѣM盆ln Ӿ߽b~魈J; ߇' 8DY˚4ߜOߔ 9ȶgztbM =˽kQݗYR{, .o~y xfss̭Ⱥyh_}c5SYj>/oNL=]٫ƴS}F5ֆ͏~԰<#4O/O{QnHx%Q?gnq2$-]忻Jb'=x0Sg6a9}oqh9uj6fd lҲdҪJZ߽{y9򊅫1]t(hUǑV 8ԅJku1:iT.`}=¾o (7m|[i3ǏN&:U<$W,XxPrСT&C#ŦWG:)OޓHжg)e8!xI)aǪAJ܊J[u NfgRVj-/xH(>l.֝*vVޡ18ER|Nj85]j@ "YZI`T!nT#˄;QPCTRTs-AӠWHŔa=N t) ^;ż9h=?J3t,],y tqZU/O﯇e|9zC8$M 9H5 tF):tR"3cD0Hsh,$xc80ZMGmyEYWs -i2O6I,}n_BAb_*;r%]۫(x~VhIYgL'%':$JIoZ+q+1KBBŐ}$͘e >hm0{vA)Dn%־9m HD+{Y27i'G\ ŽcCGWBGjeoFF+(kbr|Q:2܀\siV|*_O3) mwnwSd'Go,P ɴ#A]inݍ hw/i|7vаӜ!>Jn߫_^=bo{qP/WE'ffk?r쟓. tV1}ߋXcf9m~k'"]>E׉LEq?GtN3 X MBЯ$tt@Cp4\+ 7p Y TǭQGY$A<3 R :KL˺t*fa)S^r18-2qYVgNg+텪gQ͜ 3&V nry_c6]ލ\p2Bmb.|2fٶXN{.l~M8}{(iB٨X='57ӿ `"+͊Һ[XrF|ٮ_v}s?krw(:_d!F /8sXv;f]su^lr߭8%-Fy-ȭW]cL5.J~+ZTC5.V*vறPP|)4E"@ShyM)4OiMe)7EcM)4E0 `7E<"@ShM)4tDGRr-ŶbRl[mKm)-Ŷ~FGC,lfhPXVhZ9xb%6W! ("@ShM)4E"@ShM)4EpD)8uϳ"΍蹤I'Gc|V`U'N:;Xߞ0"V"ҩKňt:Q[Ý63Dd6IL*zk08lֈ nP ]Rd͒,)X SlL{M?B+>*nBCLO̫뽥7Mg>w`)= *p?_;$ؠϕblu3ih {0 "iD':˸h2hε?a?a?a?a?8c6 @Bkg ڥ s!a6(BABg@볣I4Ϝ`}$Yh*ݿNB;MaJYdx\gz8_!Yn>v6 ,0*H(O"RMTSZ[NW7뫮 T h Qh4QGT2i RYAZWbwHdcP.ijxabҦx PwHFhl̥ Ī.8ҋ?[pv‡:ht{aj#k)#jwiwea][:`F'?\fwjWWB 5y?= hveG?@w8&gJ2F?Qr۳Yjzo$ZqzBeJr92nqAӴ0اE?S*?WEN?~?εtpeZ,4aR8A$c$qotwz r$!0q{%n!qΔ99MLϧ"gΏ7w b@uv׽-3wk(BVr3+b 'dH#]c`01QyN>ng ={okGy$׍vzZtu`LF矃ϯJ|wNhg09}@ԆJO^Onw|o}O돟~Ƿ W`}.ٳʞL¿Fy/ 8lh.C˒.2ZrkiŸw]?[kbc痛?hRI:ݮ99C'>d_MΆ҄;UUFb?C`_|~c#FZ&W@J:@w ^EK)pƸ#A3o%Q@29 ZsEjs{f:&z! ݾQK!D%*0JG8h+D0^hv컶c xb@vv5,?8QSO96zMzzdٮjC.r)Qˎ]{%Ѽpl(\r=^#&(AsU=ͱM2Jk Ve'm/] >h<-['y Y4l^W/lK-J6nIu$J"bAb2q*Y,ZŒbyY,A8 J0'2T1WqkdelH$Z}+?g+Mxr l%)E#Bklͽ[h y&[ޔh$!KRTR;Z6ה*!p 4d`5jR ₁l"h8*x AZ_&i1r$NewZ˚/UU펵_e1W/7xsקD)g:o(%HBF"ܳ7L L` hl>Vxzdwϧ=W#[bI[e\E%x¾m͑ 7b,r; 8=;[u? "^ w ɣY835383 "D1VׅhR[9Gr\^*!`JRZlť1%`y@P| ˼Y*,UX捨ES@A\zV ȕ28RYmh%3B Z8/jjIC#VrNmTC8RGfj"/i}$ZJ% Ŕe<2$WSr*1eZ0M)T.1e/&L\8"ykx~,B4 <B#ѲFs *m[| ,yвhZ@ˎ0VF#II@<'M)xc8`:h)X֔ %gsP\$CBB82L+X!D[)cŅ­&ae?VFq~=~pd+|l$Vhu_NY8'zN0>.u=xfn!2<|'GxF7C_[]M+7==Bn }Ϭ_»ً2B;_w_mh_3`ZCޭűaIY%nGx&iLu+::0Myf\xy`ߎG[޹f:cQę&g,=2|&&7XфM~E{v@x:`GNYSiaDNZk_k$|ƳK}@91+jԞ8Nx 9N F_\6cRL}<8Izm~Kf~A'9bC[Ĺ>!oRbgc?Űx]u\-l!"0 :Nf&J UMoOl|^!'/JfTLřJ{+ANhX[V?"M.zgfm܎u룙?o_-c<2v(//kr{s )hs㔹sr)J@mr0kC#d@⍫΋fPחX!ZTd"d$I![TNR (*F:*Q(Q$wAo!ZCQӚSDpJ Cblxz}sۦb|r?֭WYm>9E'TñzpVkq>*)EBlPF @;*8EJ#e[6ܩa`A fZsD Q@bC2.: :%nU Jf)&TR5c1rkVQta1WºP2؆Mu?;Y)ֿvt;2q@e{^w45L;q]@)y+1DL%U \YmJ4+@q2 qdɤ;ɃTB UEn"72WtRln )Zw쫵MamZZ%TURYﴠ+m#,*T!/!m:TxxĻQY[G|.VZ[1c#3p,() keTG\1(2VZI;@(OjOc&cV`#+qBAEDu{\0~Va:PovzYo^͹vP b%g \2@9h c{SH2@%~޼E. 0[$P.W?zǫ_'uw'1_~o/?yI p/.eoB+dG[Bb$q"AY#sgJVtV.(%DP)Kⱃ39oTսٳj*+A ˽h*nuP: \8sߧE3Г7=yf1ĠЦD"':9P҅(1I>1^0]FL/m}-Ԥ.r?/)m4Q\1TkXzlȑD+7}j;oRKw"cr椩2^ʹubAk(Q4Z02%]_֊+?B 5WQDZhoprl+. 0krZr oy0pb;Q5$X 4"sh *߻{ uvZtc2xP2R"AqˑNmlP"EA,' X3TvPP\hª-?U2@#p1"Roϸdq"6yEe$D۾o) d@Ƥ|4*5T h Qh4QGT2i RYA'(Ơ\҉g[9 M C.!XWւgײ;77~~q33w텉|D4@G";*8I:?{F!_f{~YOf]`|:-$bwKe$ۭN'QSd5YUXj270Y|}0W9W1LܞL ^āAZh9pV%c9P+Wˁp _ʳKXƄGl7elԳ3uNa^b%c$e,ߪ 9ի96_Wb_TH>SݺW5ϧ"]/wӋEp<_o" 2s7 "`뼑gO ?qFRq$.Vt4 mabŰܣ`,fM>,܍ٿO׽эR~K6h6u`LFñ]_O`J|wJNHoX>ǠS՚N%ħ~Q=Ͽ??~˻ˇ_ށf`=(&{6 |<~~hƛ 4fh[Ÿ bܻ 5׭1wE?9,FyXful-B/̾d_UgCIo%KTiU*[ 4 ~ϯ*侍F&wq6·rCIGC7zثhq(8c᠙(a#d$sXК;-Rg(0P)>3.0k D;W !`X$V:̩BH㵡N+;GU5xb"6y=p9;GmI7d-XzmWu! 4 qZtU C]eUU:*]JWGQtuT:*]ήJWGQ®JWGQt(mu?YN숭wDUtIZWzUH`2/(w\_z2T}]g|~L0]טO)Ϊ=/@#tYua$s9G6`MPxPsdkMr+0 b@mfх-k{F'&2g-Ӛ `1$1ygLX 9.M]F7X:5Oz߆I.9mSqY o/i}gݖe9p3ؕQ9gZ h}UYmZdKHEN0e blpM 1A)FQ), qD'hѸ&k/,p+={]jp^Ot;οLTH6gd/Usʄ2bPV ,\0^4xCDrY8`,o5EThR̂P%Γ9%x%x<NSS+_z5) vEgYȲD貽L0/5+K9k*ygJmxj0l;†_ :R$oyI:Pq!$2XJM땖)Sz,Nx8Z?@yKfy)f>_}KE.$ w7ɇ83ӽyz輌3GYeL;^{W6})9LazurQήNGR5xb6*=a.d)+BCAD)z>K4Twx9<7>#,!L)ĕ PMp$xR624yh:G.(#l)#׬ZƓdQIEn| u-NjP,z̒`^{'Ї`SIolVąƂxNA@IW"g3 4vAơSt,Δ=i(VMw3,Jh6 BBkH *-1`PjWN|Ӫ`<,"ElIq`\qYmcLD"F4tA,+eDCIH#hM"&JDRFSQ-vZPMd. )6al mB3KJHzBTtԁDgNv1`}*Y J1D :#6 qQd*Mn \2^Y76R†uA*R,wro2 )JEFHddl}v+eU' Mh+{8:w܁`ܳ 9GBO(qrQ[$\ tU'\u286bâX/1Qa̙42בLzq)v ^҆$ߗCs=.N5g7mvn*KQ9v@H3 J ŤQZj!^JEPui3wi_aEHr1fd83[TG%saymW)f ,;AN8yN3c5[mYztFzJlCt~KHgJ=U%t`qn ̙CNX`"3F+£o&4ދh%Li,NjػS/`F BT\ E,zSh! l!'o${|wٓ۴|*.M?iWawWwWcZg`kue}_Xߪ'4{[@*,z -4|g2es~V!ysŧmT&vVKa+(*: וJii8)pr,% - .IN3)(3$ R1H0XvKé I:K18N :gЄC0*@&g F8-YFΖ =+Ѱ䡯s'2et9Gu l2[;o U2U|lQ ]Zxi٧S1kI2ef R󥞜ǧ'ȭQպŌ%%h26bAw3yz~ݚ=ݜ-x,킹W7t\'KHcpYs5{6eK'7Wߺ|(We-˭dk9yAƙ`%^̩q&Wrj5Gj T ,SWxj̩AՂQ/~x`)3O=O>%Skef}#)6<8$Y"(CR"8jՠ~LW;'rzYd-?U P`?ZT')IBVbdWLyi+I[7I|`9 )4o PJ&sYռS˓W|D_oj50KǸ'Q>f9XճY1åïާg &\)2sr:ɏ'֓<)fckf%%u?>8ԻJ;uvAVo{a 0v!6t&=/ x⼟/c6C4/ey1P,Kb@ءXB(&bPM@lCUN CFaDŽVjW zL ;l~eG[Bǯj;Ze脆eNMPxP8%8ZS+ 0 "%/L !{ -,6c]зn5֎=}0ovILˆ1 >$S<\1V } f'.O7maq>EhXhY0-j͎z)*S)2t7|sfoj13AFwOGpSmf8 sXu1k:1?x* -qǿ>ퟟ5´4jc _Q Θ6|EHʙrhA+|OUn'ƌ\AkhAJw*ٻFr#t g~ 0A\nXi H$SnImY/K%ٽxgiv*Xb-]0.zw} h6j69xs0ڠ 8 UF'N4tNPDKb mkI9%]FeJeۻ5`\[q>+8Lk1_)ӚHFJ|[|גl+ 6 Yk s+^6ŽF!9&v6a$VlJ5?9XwWsp]m],eT]Rf]'/&0 S+d*>``XЖ}ofBհ2jr} 'I=, Ŏ7Mfs@""{$WoJg֌eq˷da+vM0я9"2 Tgaa{a<h ^/|j-Ǣ `A 4^~[-_,p?c~,N$hNO,YɗOw2Z$8!F!&(ŨPMڄQg4YN6#g{FЉ\=>I56|@ycP;f= r^J[Oj+j)g pLW dnsTqo^^+\/`xh;΀7.MW;f3|?Mg6{nay{eT~.fo߫tY'7ܜ-r7N@~gM`<^pɀrw{D Zl _CE&J˷ wB2;tdYwFX:xcX8:x75,Ub[6^}@hDmK,M] 489i%xVm(ZnqGRVI4~WԴA0#Ӕab&7G&>G(jVCdKo3E j(&i[:6a7mẕM"oC`ūy,Tjm֙C_v:g΁ra(jҝ8Ox 9N F?\??/"^y/OX)xrQ&>FHI6!ysRKLsW&uX@oU{\qhgގ7fu[RnJԊp|<L1c[&MgN{WHl+d)+B/CA'D5|Ήפ>4~]EuMjxAC;Pd~eFW:{LSU̙AHTc'o3H(`T",bY#@S*ovIcA?L^ l JL$X:)$qk(j)J'hE ȱ2`JG$E2>Ћ&ABԃȍL09N־\"j4-4+wOb4K7ݨN4W| <9Ν#]n=z Ɍ׸g_ E@ڠX>2lA*b$%kÝ 4@FAl QՌRBr*Rz^ q9Q)qbD2K6*]3nF)'х8c_] B£ǽ#9k:EIv4MՓ_7@v0=;21cvS̔L0)Jf:O%WV,ql8GjI%wEMP pDc^ҹK]c0sŸc_m kminx9.Rc1r\Rle0@ d3Id,{z`!2$hϢ]"F: XlԖٮ+~hbFFl4bpjDK63 R)<(bQ{4YE8  kbPQyasu}")I;x*kqGG-b>_ճۣq)7[зC⾟.!->"M-ιh:3A C9~2Jp䵦VDC]ߊhA7!2Ky)C@ϴb iM܅\.bbIb.ygLX-ᔳZ2ȆYR !;BqF9ܙ7ȴ3-eNDet%IaSLwF ފ;0ǽTKńbAR$#J),jl{L`SIolVqo7$_xL!F Ú|aR()ll#gdSINw7Zg2?{Ua](gw?WwMXD-^¸ZSFJx80*)sX/V,gBS6M V3jc"0 ʐ /q#gGKIH#6m""))`Y˨; d,$ 1NF-`>Pxf*tBc;g>b+vd=!*:@sYv1`T6l(*@ErU1< yɘ2fAo muB*R,wro2 )JEFHddl}Nt}ڇv-֡ ܁%݁tGw 1WD݁͡y2} '(wt!@|6q_~~}{ɷ_ɟ '~#Xל~#W7_ajy\jgÇQiTN䢢Lb\+96g=Yyˮ֞īOc T o]nw/%3o}: w=i4w8]؆+hŸ~U՝mZEG>0[^< 3' ˟{Y=Q/Amtq{xNxzyk]뀥7y }Y~v&6 q=-)+" 5 䢶I.Q1BC.:g9jA>֍|t]v؈ Bbd:*Lv2בLzq+9]\VHbKX{bOf"o +ɗ > jje8<~/wMc9۵֮;cn1GgQj8;>:M3L+>{`(nJs(a7Q;䬣'ٻ6dW]W&b`F?%e>DI|HԈk ؖ8隞ӧDϡDjWHQ"1CIItl"xS!0Ġ 8@)9u+]@R,}bWYd% cB54h|KSlDQ(і+iW#W^ñF4G}z~O,h"R=Q/Eh<9MpՅn'JeUXx3Etbp-Ęץ`R9jTe!5R2c s6tksGwMΖ4cYфPab" o i"̳4Ѡ?;lPی**- $h>6l {➟xrI,v SqC?m X*Br*2dgiL\"?ώ!r罎ʉFHX`ʭG_SvO &Gpn L9fr't.h:ן?|ߗ?~ӻ?|`>٧?U/4JU"p||<_5`S]Sшt|?_M>~Z-m7Z?v~]?&1g˫R>)N, ׏8 Mj~9 r:azZܓ0ao Cb;"j_r#l6s #&9bC$ ě :9Q ϢcIqo![$hFP Z}<]8I6>֍#:}Ƌc\)tčNeNy&8)i B{(9}WbzfCcsf{Va =OZLvXÙa&wLͱ)*81EJ10w p\\:**{,pU5H٦zpQGW$hઈ+ձU>p JpL\9*⚣1IZH WWv&Nj\spZ}&/ *&J|˰?t~`&#GSH+~ϮHٲ W‚4fX&:oo;8e7.x˔h+*t$&y1 -E%ڜxJFJα̵H߱ s(p?WA|a=*'zj27FE/ a :ѯ_n*wcG?{H1|'s=ك_({>TsZgV[UGndlMD\JJf0 nS^p18-"q,F3YtVz=:O[;O|B$W߉v f`DRIqɄ@4NsA %41q6#׋D.r^gcd[\ "oqŵ3[Z\| \<˹ٱ-Bx @!qEܛ8]kjk1~ n8q37>+M帟T)wK]wާ!IimY c!Dxxk;F-y"Lm ƀɂp\h5^c'5Pju6ò9Ľsh e61xFO[3c!\dU y⨗ѧgEZ1 Ћ5 ~gLʲ?Yt:wM\}~U~n~WP"z*͹*.YqkX|B-pɋZ,5zRR=Q*nq1PmgKmY7ȨBB"eEFAI.3<"ŀ.@Î0q8W}ԭ ,7iB5J=ε`_Bsp,W+4ܵ>)wEeРrZ%DP)`ΐPl-_(5 ]6!VyQީ-.ѾuW .6Pl6awvcv;fO1B7?:_̠WJ&<4JԊ%tPVAYr Q3hʣ&;ښù}QkRۗ,x6nуR(TdiȂYMTډ%y&>7.%DI=X"#ғ%im\զX[cLསf戍#BWwaű|EBRUwͼNvu1ѽzzv|sOSJkM{]%fjtUtaƃϣK~1Y "~+e(5hZBS8wS Q:UҲQ9im.lzNJ6.cxJ 4A2Q_ڢ9XN@r7sنgc<2#лP] ۓOZO/S!>v4eٚ]is+w=Y؀%{_vw `IfW+2(tEJ4m-NIֲ&b5{B,,':`'3' `rr\k)U>Yo@8 o~nQl^nSLWe'RkΚ|W f`yf+.uL*D х֎|`pu*^ B 㹵=|j };LAo;y}eA_$֋ۛ:51#2ƳvMNڹһOM%'YC[U,k՛ԛsP}ޔU BYiJW˶@ˈxX<-[Zbm+$մݸz wK4UV1,VCWU+P݅iŔ~IJSMB=iIm6wfa'j6zx Ps(wH(wo Eez; e6 $MrFtBkfDE.Gk$.' 7ߝGy=lu+MHӊ1b+hjP=ٻ6r$4ewp6!d7  YJ$+v^[R- `;nEvXU߬1xoxSTʸfUConWt4ݎ&wU6DDI־վs(&فr`Z88?l#d46MKc5_ʭUvk M P$`n  ! 7^Fɴ$F:e85Gkl\rc]yqi[ ݁]@z@V`f^(̥bEEq* RJ/mPqxqc4p8Q$B ꃔcw*PlQ'`"y-ʣ>39H?Ht|tp"K}b#*>REc֎ViG#,xsWL#,8ZIwš-a* k}/C1$}+)oC]_.àsi{UqQ,&}ݻW_?/USqL ~ nL N:T*־'=A堤?Bf_.8n U֨xʹ-jxY{&U/AVgn"LE. m 7R?nG憩6-r 2(Pz]ҁYO-3Sv_wNkF7 V.eaI{h6.k{zgT{J64n>RWZԍ6w~D鋚-j ߙ܌ZBo:sQ xy,n,n7 h>]^`1I_ؤ Ok()Ld@pAlrfXKh+wGV"Xo`8PqD`bTj[%(iނf3jX$"﵌ `l5ͭ,:#ghZ68Kzk7Lr}G6ޥ42C]m*6j3[4욦]xדMNjw\IZ](QW^dOj@LE5+etK6+:jefWMvwQg“!%7Iܰ)GGEd`,3r؂Z ?D `(+22uH+ޔ;ELsoYy&Շ+XO%A]ڼ=ލ3@aD(-udR;%Jk\TJ|Bڻ Gz/'ƞzj (9D)9Ĵ&D P*bLu<3t h8ō8`1waʽ3ø)acU3r6:>iz; pru=ՄƳ_O瓛z[wCׅN-+0uHgW%=N5vT. üKF(֬kkF=W;ľtdu2 |u>>HQK'Z 3kBV$HF/ ȸČsqc-H<3{/k lUqNh C3, `BX>D V"QxJuVE9޹z\-#)<2*T2UI 8XAPVdb^.:h "`7*E fW $n{j놴wIk7a%`,|8N'z={m?|wsLxv LK50Htgzu_7Ю]S6Zt9f7.˾ʠZ- J?|}=oz4:IN]> 8 |_a_)ֽq&*v*S|%B)m~4{0̖m6Ҳ6}d <`CIM@o 'A"3Ҡ @aIԂZBT+Ŭ1JOqʼZ>H9a=5[e-5D26hdE(ùա$trWĖ] n{8ΏQ>Y-e*0=K \py@LD&R H{D%r \F -W\v0*QվDȅ_ W`F\%r:qԦ{.eqcCW@0cc &r9=qdUr0#8K&$`D."tﵫJ WID0?U"W*Q^JTjKW >gЅOr1&8[8 NAg1z #%a)!Wb4ԏz>;\O:Zʈ`oo?n?n&~_{16o_g'<oɇZ`8xr^5m &cu?4.h\[n>F\FEFNa0 ,RJYs>54k9oV*߼:^r-S TRėZrV: 0wz|Ay/y r-o| 2ס zMN,+yK,ՌMUGTRJmɌ6>fl*d'c%r=&jwK.QQ^%8Eb<`D.=?yV}WJO/R\ iNmO snKk|_a3./"Ny?GN JVR9 60*X^ǚ7fbw\j]DָIGje,H!kb1UPk vs4QwK)( |R9܋Tz#!֖<2AGBS獃i0q)d*=6Ba:63( &1eG H{WJSYT񧣢JьGM}WG[L.ѱ@#+C`e !2VX/"W#X+C`e !2VXkOUL03s4j,`K=+e ?2Xc~,Ïe\ 9Ïe~,Ïe ?2Xc~,Ïe ?2Xc~,Ïe ?2X{14E"XЃ)MȕSI]hJBriX0ufrѷ.B0#gR!JLi)1H_眭]rS娳QgO4Wqy$#IGTţ{:N݆a"{&CDTM)RR0DpCLDK9 Q u9:~m74fDhpۿS+o| ܐNc00+9(P H(j5EJkO%Daorma׌\]85gKfttҁuQu6]ѩj-lZ 뛹+Kd8Zۤٻ6$+*>9!)RFP~zHQ("(P"JxWexxO٤ƗkHnYsI6Ƹ=/Fx%}b;CA܁M|;݄냩7۫y?oԾ'"{>j-Oh..y?(|t ˲8HO٢GKk> ]rrV&im=5Qu.*nFD]\FKah;ΣdsrŇb8d鱎\/?|tdpb)~A avI{(wya \sc&4̸U6T3F+@tRU=Q ?A -xpģ5ǯ%pɺ8M(2L66)eXxdKU0 K>bB1qU6F 2k*~b1*͝B$&Yu4JĤ!_ٯ?ڿdCUlu*,Jv>Bn7!J!w zM5wsԬ6E5GIl10ت#)̇5i79d{r5njGYG7(km "fܐQU^6Z:ie{E`Q`,SaXEc;gP(! c:QSTZ(Ui0Vҩ+i `dFj3Zzk# ho*퀄M٤a |4|Vsr7ǥ F˲z'_0V~5-+933Ɏ7J*%nn8cf=+[{ pQa>;Z,:u,kߵ昴5g5rFj n0S8Xҷ 3fLX0$:ةA*Wsw,6ZQz BJdj*I hHyx@uX71v}u [QW Ժ) <@O?!o ulk 0iQH0RJ#x[z#4~ ڨbj#Ǥze#1$6gVE2R7kj*H]6 |t~g҃d&S` VWXB%jϠhd?샩_GӰX* Y;0m@ "֚&#-m.o0ȅg lڼ?-'/#%ق9>X2'`l9 Lֱ w ҭ.<+p4TZ66:SS/6Kcb`XKEu,* ZIgJ8 ؼ@좀k5#Z`4rM3p]OvUp@EJ+e ^v/n0B,<1LYt+pĸ׽d*ם X%ڊR++`F3v-=ܗDϺ+[yסtsz5J6+EwT:*=]]gY2_k\}>=@ wWۼ"*+3V-p:̠}t3h͠AK'1H W9[@r8CJגo mѕ2v2[|9ǻőWd_o)ˣc+.?ۘqGFOY{Xn};5[!/lT Fͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥ\ͥm6'<i6d_M9o 'l6Iolf2~}[FXbB#k=I%iT\V0cBLYџݝx#:\#;FtwE٬rM\f}%/`JHyѥdT-;op,hЦ3(FioD߼"*ȿR{z:ۏWgwF|ٛ]#wz }sc %}>s}tݍ[l֮]RY@ E:{M)fbiUF#x9#x/z,hI~j֑: b-qW iLD? ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј ј јu5&sѽ&Lm^uhL`N^cV$DDz՝_7]75e(zs]@tί~UE6y3)JW(u~7.6h _Wo/߽zq}f2X~\VެnK4q{ۦM"?"?M9._{7m+]g~"}._>gezub<ۈov$kjs%7~Հ$n$nN2nfhp 7o fk'mZ8i8-§ st|-qy:q{"SS]/X?9(n IL|~hQv3Q%u6l>+ ā0rR8w} |Z_?dSrKOꢫG<σ$NIۭImsIcV,"~f,^MؕX+'w>~vv>roy+qlg6^Ū]XQV#'1a;#Xo$7xr);sҝbF@'v+wSCٻ C+CQky|;ɄgA\>|8e{ ,^&u1*2:L7-;l$_mgK|:o?Vm$^?Ƅ|+P"*WN0?{NpԦomxE4Icix{Wn)cA =N͌3fǑ9aipg<5ɹ0J.|vQٯrן/Hm?+H.ve]ˋ=`"'퇥KMc(}# Gfޔj F!vjƶ20 bi}*O-ER.!NYa_vm/!;= y$؇`Yr$x&Sl],jI#SRk̨iv"_ڭhVL̮hbPm kmSvEVќk)Q`QcD`I.)2@ d3NƢާwZ2 Cs0( tMK#(%`TGema}XL1Y1vE#[jDjyU#Vز1x P,c \{x6ߜ l\OāRƗvH/[#>FɢB|Vɡz֋ŪF U`ǤD)Y4v N :ZQZ1bg,RP}H Cz>< QhCܜ;q p;Ƴ(ӟ(=*SJ$<T<@fxЏ0LWP`Se%ѶSΉvmi%C}/I1HdlH('D< oGm\Hߺ\XU~\eUem_vt=]`|akHLHKP-ѐlh(kΐ?.!-saM7PH  O{lΠD5 N*9T͠ W!]ֲ$atL[qդ`1Ȝ #r3pJ Seju;T0z|__MU8?fwt=?)[E]fЃ@h{׹;=12}Akhz{0ɎNmnl6BK-TvA{rzF[[U[b9BNxi`+% bz ANNq$Wm1 &M$\%86bâX/1Qa̙42בLzq)XQ+p3Y=WH]Y%Ӗ둯69j5P'S_}ez8VtttDXdgS{#՚!nC:pNMDDTHɓA"X1("$׳2s'3ZxkbbJU6D8!s-%Ht I됶M`e>zcM`oO GrgBA }}RGHo"ZTh(VC;"cH9O$!DeD:V2x2\țPțD2W$$vYʜCU61+%Z*qJ9_Dh#hQ(XLA(XL=c3%hCZzW!R[9m0ۚP"Ao\03*xLx*r+r1\J),]/RRlj)7Xʁ0N$;Nֲe"f1WA5S. Q.V9/s3%]Ɏ']ɮY)E_M|u* R.j>DP<@ sKa:bMMN8MU R&9N\"kM=:Wj%,yͣ9Mvjuײ֣W pAKra-O sb;ajd`1Wt޳&TW <)q&1%G==H! }}A˳@dnt]4Ԝ]|{mb6 '=?ϤҘ'}hT qh Zkacl^SX\@J: ;^E)pƸA3oQ#d$sXК;-R5NqXݕHK3Hc!ǨE`oшPV=>/6B[/8f-Sgle'%&. KLRI1H-xb҇UL8 !&(ŨP\0*6a7g4YN/MWLw|ǣnm1RZv 1AI$<4lR9?+»^ESL#(B2*8,zLV(n"w޾h~Ɏପ Nڹb̛ 7go[ܾ Ɍ3Mr zWL0% A624YL 'A WcQwSwQwQwNPwm0G <%Ep̍"9>6ޏ!N C@+!!i4gZJi[8䝃CSt~:QZm8&-\d)o Ŏ:2艀g$VԱ %gsH<&h]0NeTYrι,K 8{ƍa~?ͅx9f.RH7z%D8Yq@NOd(!R<7\PBFXi+RLfӁEApBoHN;k@O[V[\$.iOu(E)H,l-{oY>FԖ|vm[|bS4Zg,yzA^L1W`z)tCZ s0]64D+Y4GJ)M9D !R"ևln#T^nY' Nj,QqQsP'DPMae5G#JhvJyCLZhQAIҼk.a>hYE~@xe0gƦRD;K{0¤B:O)9ƤPFJX~Rn% ǪZ/b8bX <7$z8(1&E)$II E.6kxO@λK@Aw{ktEQ9FY3htl9гzސDE VHl椥#/8OD'q ($eTAB]?/!Hր譾(LLj\ymCd^@&q=!6q8]{Yj7F}jfw5C!Cۄz)]R_ŀFoA4|IƥI.qbd*,DJ lp,+wKD ]d&X!hSnZ0`yI(Dp'3-\]@w~3_m6A2c椌,B7A=!Q7a~wj6o6{ӕ=riX+B2bhh+ ) Z`fUx_Ⱦ}J{6(D$:2dVݖ= b*C0.W']d{> Ujg@"] {U}>GC}|g8~FO]G`.ۼ<5i:BthFHa1A5cJV E.Bџ|F{r|F9dc>)cؘ221hE0&;BשdUR0S|༷! YVg.r.{g艏Lx㱆'!Okbu }xLLQf,G 鞗襳%*Xrp5mL[r, ې0A#/ {$=ɕf&T6jl7t^b43iVSܬͺ鋗ZMOK,f7m oCI>|OﰒN.ڝn훷ܺީ+ް7*"B /ӰJ:z0*DJ XuHm!Rݍ et1{ : )=)`Pl}<MWqeXMXTj/Xh, RLH"o*¨V2Of) 2u-^ز me'7ZLrt>AeIƆ|QPΘFJO5^I˭PΝlrh-oVoq\bV$T WLBYZ[! & SY "EWo\+c8Lh,qdz V(~ ڨ"QѢ|r獯}8;yH}J-kЧ';\1+G}[Bhgݡ˝믹^FzH /:9mؑ Ψ#&qZ`S+o z"4AmU&8$ˡ&qJ20xɀ%^XҾ,is'j)K.Q ]lԸJ.u]FkRiE\PYQD63GE42J5LE)&)^p[io趧跫qlQާxZf8&+s5 1k'oS'G}4icdz= WdfXW8@xȊI.Zf3<'Bqd'ߗ,כ{ݮݪڊRUP3@ވB8+*8(rSnq܁ ;oFa"E@[>+3$EɜbgFV ȹB'2j8efܺwu,wB>)ڲ^tmOH rxOjZU*ev+I7Uڷ}{qh;|Zf[]^/t~I̳GN'k}zU]#C6'ߵV-_~?}>&{|s!~*iI⋃AXF -D\㇟?_e /i,.S2/̸W2qb_&ᇕvS")9yZ])&HO'"至Ss :N{>ڷ,kz[vLsɷj-tbCvwu6yQhy/rZ]-WK;d3zǟ-̕7e}yGlNFV˳QLJ=+ \UzWEJ!zpX GW(,v4pU]3UEHW$AhઈͱW$%fW洶\"P6BܩrE͟~\[kox19bxͦeRK~L}N#g?{ť8"&%GE\85"trmfm"c"+k-*j},pUpJ9w W )/K`p/Zy18K#7FWn\:{1ޝE+V?h%4' 31^4>".Y+L0]a]´b2k}4w觿&w _TjOAq#G^1O{p|q淅l4])$A68x{)JGhF78Ih!dTuo_%A_Cj)Mbw~ؒ'}9%z9!^Q"Xri2v01]FQ9j$ d9&&gPkR, I" 3x!0|ԖhYWgzާ;.t&awGwE*K( zɬHa6,֐* fkm+G,0/sڼ_lg ,E"KIN^dˎn(89ICU},VSVs;-Z͛vo'onfޮ5 Fy8ϭsQD6f֖|{ џcF|HJj0b0\kYUd1-F Y,+x4]N~`kGl˗4wuɻQIZHXeW8FN Fg8$uOo|S>.܇woh#,aoI.o ૡ/-diҩ}65z1{+yW0kƽǾ֡r(8>};.ۚ1TI^񠐮ףxBb_ᢋV<1«B#qش/ܞm(Gz&V+% L]0 QIN"X`EF!\2:ӹw{ayMk>wq*6{{NbhB1D*yC9 fS L`J8ڐuxt^;Uvivs}/6ăVaW'sz:x8uNk ᣗ.5ʇ$]:_+)Rٰ|ǣ/Ü+D.Y57$4x4qLl\ǟ'1M׆2=9h`/{qجllr7bE$:e<.(2a|-У2%G+2+i8g0@rrUHDh)u g*Ep<oqrQ,YT !l=mJs".5 kdԺekGO1_b*T}yxIKL3ptjM~i[A-wS}DY F6SSGM1MuF N0<֌Eə}j9Af˜&OOf#Ϸ<.hJ;! `%g"8I܂)ʠ|Ey OâG/ A#/R[Τ) `7Rg,HB娲KA;u1Y}'u[c}H-6q(dhPO"E=A- 5,!+ ^ x.A34a\#otO#4(!l=9&]`%St 0^ >-i\.8BEɬW-N]]@wz;wH6h([M}o))EX,D[$Z,+wNF}>ls7]*"ТEѠ 2† ZڭwKvtFm~ƻMjm v@K±ɺS Б"D CbKi\o_x4osMІyb)egm2ĥ7])ziWlݫϳEi_d[z<3Z]} U?_{:?-.UwP0$hEӇw6/?|7-Z^ugh;h&6? ĹrϴEt@̍7Jz8Ѥ)蔼=]ew$ >j$0sMYJ%KD)ZSrN{eJ1H6JFg+I>SAq"B<$!er/@0jy k+dwiRƼsz1%bRg]nJ16{AK  ZQ&I1E Y@tv~ּT #Bj&e+j5qvj(nT3!/:M/LJެۮlUӓ׾6MnFOW!9Xrg<@S7]㛯UQyj7Zʗe$$52ĥAȜWFdY@{j1@#6AOBNۘ}(YMp#s HFjFz\VӌCUBcBf7o8?l_vWNqs40?7&K\@܅IkrFO^V.'C̈"X,-Yɬh AHQ`SF>L" &c _á8;Gq jWӎCQWFmߣv`iwJRʌGFDDb|ҷun*$"!WEL&|" c*H%Cr)[~V+XM?DEb5D<4GȘK-`ngʨmbLpDAUD$|㤕$b,* ΄ЩdV&HPV ֮[M݈2W.]묦%"F kWV!Dv|΂Y3X ZK,@8EHQ'ܪExεP<P'w8R5rn810yэAp} k{]/%bDSZ1UidV t &B- 'WdRŖvO)ǰ3iy9u]*K<H8GBt8,C"H?a>梌0;$ nbjť]GJ^8zbzHgqxFtt6veYLS˻~8;{4->^7}"WdzKiO^RzJEvyf͗VQ2e}f24 i<ܔѐ# lϣ V~ҪճedYծ']Hc82ҵL녘04ɘփ/|ͧVZI1@6A h gOQdKqWqۭGYkk;`&hQr 7\,;^w]/^ȭͫk(DhwnA0ͲjJi[W_Us5F-{^i]&OwY:Wֆ|Gx;]ᬕ..\Z|ߢAWo]WOo=7!\svQfϛy*c}K]ca$%DotDp&c6lEdz 9tȳ5<ۗ6l \aIy̌r;|oN8c2<1MUgYgȬ" Xr}P9Ӽ$)Sn|kHQv5qȖ7 oY\ÏgVIHoKo`qCl]\rwQ:` 5RvNhcSp/2pFJ*UmR;h,Je[S28*$DCY}ۺ8{nL}MAzK鎥^u*̝gl< OHe6`iK!Ih(},!tC'KuL uˀD .g.G,h!-wP<خZh#얍ƀpS;o *sY$CI0c0PDLFGVEgJȕ"˛YLۼY ,IfLLf) %Zvy}ݭÇ[m[N:@DMEv4v{l$Wk?m( mecPg|"sʰKEΣx_HٟQ@k+0ESB QFQ"<].Bql$GD}_Q"26ेȩsV%mPE p˸|br6#;ku3#Kփ&?xN,!< _O1n9yء'BGpxLaFQSdx!M_T$S6"RJS8Le,5Jncdg_)=6*B,a ܙ B="$74O+NQX VIWp#[0GD 3?/<ߗ:Ͳ-#-N=qqMq^xO3xǓ@6®˃:-7yy?ygO bOǓ 0c|w#WG\ߏFK F%z/*HYOIg9u9} / 6*Td.ݶCWن^2lq_ɆF-p՝h|Dlً=3=Zz?A稟z ^^'-{ȍdNj.~ryk@_=HCvḟ_;n%>>&cR\BRD,:O7lI9WP}!T!J>p>$n<}9"8|J*N4#HtHިBX g+HŸNyt]tnX ؞ h xb1h ad̡"s!$3 ,g$2xk7KIqXi=76t"j$gSG bI_K<ھW-r6{&K}L3PqN(3϶\(qOݵPwrO䌸'RKpsӏ(>cɈy~?bJ6Uۀ _i%/B}WaM;}*헃ji~,rA!~J*"؂JKWR]i8m؁jP< Qҽ-No%9/ yxPs@u!HULy1!ػKG)f/ E.NrP!gim wa19ȗ#F%d62e&EğrL@uZc NM)q)W4.Y *T"0P,J'AuJ`:oxH*{D3l %H"o혗&lJN` ixh1qfc^)¤˝ co_.Y\?Շ*QО; Cmja}Fm.E*;8(Iz?PY~m0e_&ٓp73n#OϽkd9 \R@e OvxFZqN?eO򰜢<=CS4fAhrVN<].Ԟh+Ȝp 2MyEq4S<>=ůKJ4+rr}/^,G5?CU+rmSˁ(J̧6{g !qrRuܻ2un͌\g]28bO( zmY풡YmĎ;U#ir$!ot0l0u[YfOC++W1rb`nrVQg\5ʹ>8#y`xs%ʯ#zH{qCa5*;=?8;Av_:~ͻW|׻Wo?Sǯw4\j LŸ`v пcokho64ЪY7 Ÿ9㾿e܅kNgk8۫Qx3(^OrѶNS ٸO*.*~FaTP@ q? .6{VԉBX#]Ew+"GAg(A4 q3wFx/S Gpo%QΰdwA<2uc[f2hux ަFEr[\D~K 1 x \N&R*ƃacNg;ʂw56W dkSòǶN=Sw|_sWu1LO$@sz 6QvaڦJx@{~-{]n|Yj O4@EPH$9>e pZzb 9V*H$Us<(%8x:L"& -D8BO.J9GfBFj(ODL5r&΢?}:ar֠4[o*92مJ:om\Q/m%5lye!='B%QH!JU?0RMb'D"d^nظ `x@Zpc% / 4% p,iFO1%ESj>ʇ tL$Hj֜Iܜ)i;GR9KVP.O9[$: p[2ҋ|#9[q3Ob}7yk&5X1E]AHY`*3QJZ}Y^FP)ZR"OO;g9N{"X >L{)iEdn0}ڍO-%qĒԚb i0z>8L DNFC0ۛM_m֫bŞ2Dԝ@Qۚm&-6xM]{߹~N%ƯH`q÷k|u{㳳:I$ Ϲ#qA?;)آ?>; M5鏣;y,iڃ&t|a+><ʐ՗$ݍ  uƅ_}u+ߤ{ j1ĕTKzkHۈe>[Yկ;jp.wi紖>#>6̭}xD3//i< )6hp-}y-k(jt֛{frCW8~j!wآ($J` +XHj :1s%۷aҮ{~.mR ="6eI\/ !:hLHz)4$+ƍmJ&PcpruP)$!(OHxokl9Vd ߿zKh?돡RC"Ge9(ٓNXOe\Iۍ hrY\p$LߒhGH8$qUN5Fm(&K$S}R[箥*cE-{X''qF[SA3^a*7;*A%|>U(Ҋ)WRO ؆3$b :# -h2=?uHƱ`c1(T PH-[#gdl& [BhYB' \ctU򪿥0Ty`4 T7JY M U8Qƿ}-/^▵Uo4.Hg ;GU9˲"+ĕ{{f h=CϦZ{e]2;ejfóA2A̮_ɠt,)-W.T -E)Iu4w].Ɠ@]GEJ!5 xG)prP=c/%ӭ|Ҁ\4F.4~5Axt\&H,-5g&Ύ&)H?S3S<+V}2affqrm slrgŘ*(yd WF(^ sP|IUHImʾgv1E1Eedy+@ݒdz; ݚc1)U  7O:U*xɛRɨV2KP opVܲoi;Pj.,L^ x^ R0s$D&CdJHadҜfν枝5tX;M:a.)5)pi\P!YQĄs:g&EKZTr죆5mbפg^9&;ºֿ1}v+ʚn[ex;=ӜCvquJ&363L {i@==zzT)I;άQG`D\IVLqʑ2&,׉DhףJI f >eB$huf(YJ{P-o&ΎǫMoOϦw}f}*Krhqfr /fzb!yK,]&uMy ޛ $v[{/orY敒Cq ҲͻyԐvX&d}ōOϝ0nzbhΛ?/eȆwɲ[Qܜ]~O۽-|&=gzV7ԓlO&%neRO4LS0i ?*js(pUpRj{+0Ju@pFR`HUH)YWZ+Py0pUePH;ϮpJXFA\:*Z{(pU3u*Rr;+iՎkfp0pUgWEJ)zzpPrRgr> ߛLǕnŴ9b@䧔~n0{syF cpJtoiXtWC"]"Y{imעtp 5]"bW[`YP5+7a7+jx|V+G+積+?l&u KbR$D 1sQN1Vwjvjer; ?W?Om':8- d`/k~\, +PD;&'շ - 9>,"\4kGZ~3gLş?.Lh=]VooK::sbm8>C Jbz'*@>'Z,8Q` xg"8U0+zT95S`dkn5CH])|!ii׍#?EdEt<Yn^1J T4 s`K}׷ɗ=dgY{I.ȝ z1YPƔEGšxܰ\6shP]%|1 l@}V2p"$eZǏ>:!Sc7`Ond)V`$S8T,QKRE4ˌSID&w-m2K%/XP) )3RFAJ1L,'RiSvmߐj`٥}6Y=iSE2whGNYLE2s-HP '</?F">T'b XvDl0RNH2Sn _J2W.?=-j)x?+ΧeM͙8f9EQ\}i9Mo347Q49#X3(M$T^@I+ppх;"|n?QF3cӢw~;7jK\ H^ž0jxspAQ l8=;*iEE&k\+q IK*U[vO65u͈V׳Ur0岟/׋Q^}oW6pz=r}#Q~/J{kEiC%@4rwsrp9dA(/G5$&JCJw@sB F:En^Y.9φD*ھi+_Wo8b>TZ:f[ l6'q)C^yhV RW~ ~ʎO9);?e7#%1;OF )o-#":*g&^6 L"&VNT޸ `vY - ?ԁ./7]r+tx ug+,)ժx܇ .gM_ӫRvᚤDDW:ʥ*%HD`n7H29,PKVz0`%I֌P'PX]Wo@<;g}bu)jNQ3 g, KɏK1]#w#Hq#|qyEAyެq-q$'r&gNhiL i 1isVT,77׋R@t/c4t1@&sQn3<9ӡkkoE_J$Ma^ 19*1ɈB:ǹ:(CC?o&,vMbob\nqoz  杷Z(xaiF#pHQr  KQSLIѲTQ>c"FYj֜ITNΔzvT8L T1/4\XmBNzZ]G 6J3. ! F<+p[qo~l60"E V|I(\-n@Ns|P696pܦ6擋&TEO}уٿtr>ljԶ:4ltv J5F7iBmb; 6Jt'p~v^ 9y s:Yε}OILm'W$^MҸFwOa݃oΫEjho::^cp e~%;% ơ 8G-A,|N֦nL,|r5MPhc}ta~]a,(mbDnaۨV{AwiCݵq%p^׾guiJXFY6Ⱥz>uùة%)K{t)ꖔ]b@&WߐafZ0yhIՔ.ԁT-[^7HRkbٖ:9^O`,P["r"ϚjKm޷Ten֘Ѭ&fVW)n64muO;e+sAH;r4 OO6/<\k돡NHE,D1$9`'0ZGE5yeʤ`gQZ ɠMQdCz-I=ϛBOпYzOEpx@9RBXd$9F3ЧӪSH͍=._*T .켭-R| 387/GSAh¨xVi(+f2^JRk1ZcɺvXK#R ="6RHy.DB!SFi7&\u;1R21CLcpr<8g)ʃN@!9tǘ* xŻ&K{|eTR8"G2rdGSRsB:6[c9,BQEhrJ@PQ&I!HC%ьKn&VblWA",|*+,V2 !ӋQ m޹*U'Z@Fw{]|7e3K 3'aV}9:rGE,PpXb#@w -睵Ax'8RPތU\,V$4P.FOx0r_A~!YwzwfWy0ϞN5fz]gEx^XvB-QE|+њ pcH 4Ϋ/2_GqTHa\%BHU )x&qԷVNo\XU_=QVy/ t-/=z&hHK9Yiu5؊I v2DJD *E _8p8,r> tUo4%XCD%ϸמS4A 9c4BIjN G0ReHMf.qx tkbbwWk7?W̿-D[Z%$r!<"$!+ Xlqfhd|[ݹ?i{S|^sp ][={ެ=y&~c8#ȑ`0ځQ2+q@3"^ƞ!dvȴ8QC5bG ,ɝ$ "FEM hVDP=P:rDЇL+2KQv /|]7Ό evK*{(Z2KW]@Bۣd(16)tՔ.3g4.kGQc\}T]y5/<ɆOis~ 3Գ|Kd/qR365_\@4=D?5s3qs+.N O|^5A&VWk]dVQ`R1-x0p&;lU*lُlssT>JRLyQZ8fhib+N(%*J=I65s[ 4{넥9D 1C$y1Bb`0.&=GMn1isQo?YR%ziиAǻaMlmջ\R5 KJ:yS-P˙bQ ;!@05&# oMpg ǹ*D&F0肶4KBqc15;h>106Q $*ш 38q0yIQj8Kh Ml|i„CXz t1c$GRI)1Mڧwmmɞ ЮʃW*هupr*qM I9v~i ID]G)U-p_7?ТqF.)Wh&߆at̼psd7dP"x8d`Z"3h&F8K:TultlB+VuE l?V] rj' vH`iUڟHÙ戏V9@z?hzch]H )t6J1h|B2rYkӕNK e* YSdSR#y; $l\ЖeeEtV_ti8|\$FԄrȓUL9sPYFq c{=5i;K=~1f_2» _@=@Cq{;q7/}Vm) q 2W$VbgUԮ"anVTs͕\)"q9ӻbRh+c+*sU5lWIlURTgE+/2B Nc04޼}7K;[o)|92\8Og7'7LQJ8{CG{nNA\v&nY5P30(ϢK9zFti~z@,sq 2~p&$(Mh7o5kjU^"D]RFv }phas811>;ĺ_>~ǯu} *UvʮTٕ*ReWJ]+UvʮTٕ*ReWJd97 U\ΤUqݕ"-mO(Rꚴ*4GiMeٿ2Wf_+eR+xeٿ2W]%^HDYkHʮTٕ*ReWJ]acЗ^H 7)w߶`{A0ƣHs8Ϳ u@@8JY2-j3iW!u1lF90?"w"X!A>v_jE2|tm#OQAdlQsbmcڭK-JíMGc;9cJSKұ ,¾)F^g=OG u^H}TNAYe v&-3M' $ԟboHfEmjpm0ZuEՙ~_w– '«(C4wCz09%>74mFaQOߙy2 VZ)T3 t n}]Z:K$Tڜq Qtzm, z5/z[T_߿Z̬RF9 P^]$Ɖ+jF0BjŲE2y^7Sx=-pͰ<ѝ{^"JA#Ep*k0BhFh(eGKhEQCA,?ܠ^,,u $bb\rqJlɑ"$-hQZtPbq:L9W+U|a $L 8aώ!r07WrJD0M{?M8%}y$ޅ妢 ۣ?{rY\i VYXx|]p&Kɤ0=n( LpF;a{F&t&S2{i_YrfV %_JqM.v&nF/>͋>Yѩgo科1((6ds;q!%z?,~[|gu-Gu͈͠Xd@p(z-|4|[yΡ]۪`׷Vr+ZБ0hrN~^*{7 ^cP54'~=?]MǿӇO7>|~:?U/:|z.5?߾i!j[4-MӺ^_-ڕ.i۾֡퇳U )?~>f'3I9(͠,,4J7 Mj~|V h,Sx.U8 !Ņ$5gsǸGZEw;;J6 L]`c!@JJY28XdGQK/MRFk5Q 6=o) G\[(7$uffL*tF s3%Ф,$MOO8}|`Cgsb%njXެc;N=7<;}踊jГN^Q 6_iʠȜ2mERFs& /!Z^y"Xeied:! ;Z]Q p1Dq{pqIEK9Fc$RhpeTAwkĹ&M+^˚=WqoUbH=3@]=%Ǜ.䞈BWܷrB@$ 2siXF(H \),.*o';Goτ{uޛ6_V5bZA+W<:f[]|cvIZxcvK}ݗs̮2y8&'3Q6(!A #Ayt҈c"7ѩ E42>rAu jV0 Hѣd0)ն2+jq5gLnQ"/mUhZs?G4X9[|k}g 𠝍( )g6,̂`H>4k)nXQ,Bi\ C5 O`Ũy#@@$!HK !0Fp9yE - }\m;k*j LlEwZAw\yj3o`3eHwŐD+|qﳝy9(Uh'Ɠ{Fe_I]]."pXEL|ױd#'v!&񌻫]s~A(ϵ)0ܤP94C"a. Njc7ǡ2VCSՄOehT~XTG<4W{ۅjWK*d+ШSP}Uk~???Vyիj_ 0)0gFRӞ ˖k7~>ݝ|Ud|9&s \,lM{V=HU>d?%xCrq(RѸ XAtNk ߏ9"dI}2n~I-ݎW?WTOݥÔƌ@\3 x?Q-~tx?Z3w)<^3Hm?-\A[{V/f`ߌxx4*D!3L4U !;͚+Dd_MqųIZ[J Ҕ!A$(4%ej%X Fs$pmP KZ21H+^MC\/vngQyÜ6yٙ9-vNR)/@4Rńy3L/aAi 7D?CGƒU{͒RgOqBg_Mz_^BdGe]4QA+DBhsF5ɕ!X`ymk^h1'zᔰ =r\jHlEL>R(cA`-HcB JqMRpFYag2BZGJu]Caj3j {-#chnA#a1p6L{e nHYibLr))"$b<\Y-3:0L !DJl9 QB I ꟰XD|0āz,$S{+%RaBm ZCK0S0J=-1iU3АA2 hG^rx|`9P4MWn PQ1 Gl ?UOcq #H6 sja( aQϒfm:a L!4A@HHq "$^2:łKsI&\cg8* ph?~Hưd[拸-_|P7 v&#AhUȂXplPĒYLZx0sSر-?$ C? [U&zwZkp?2쳲>~|GM0E?z7dBΓ?BX>^rYΜQ[/9˝R 3rZq*S}26٥Gw+q)TeF۴[>Ͷ]}y6?7 m}U!%<{\;1,f*-52&ðݠrNEݠ j@] D4 \QXgAguqTyW!ltNJau%H;][OB<[Eիww+ٶ|=態 $*H>-dZ[pJM'iV;ɰ?* L}VZ&E`{[- (ln@HԔ1тFQ4`pH~ZW>n~Сknu{ w\ץ5iNBß  ܄^ ]7~٘z awgȄ6`BgvQvC|c+S'wMle[%c+ mlˉ4@ZUԃ%,F40y)N kS䵵N$YOW*Ik+I%|Wv}D=9P~7agC[5uO4܎1C* ZWDFy. KB`(2.1H)26 0˥&LQc(,<@:k䛮8zgx~@ *zAHt]2DΗ)t5="y=ZJtƛԡKFǚc3;rJONc(qEC@v2D (P(< )B:GA*@$p>9j2 EF*K%' 2SiƝVc  B' L}= ĭt/ZqXS&菥$K¥BpZ󀑍NOhB&:#E B-u5mDH# {C.(= ;`(AW 02q[!ET`ZiKH6u$8्*$}KKsab,x1'sM=EWFJe7@JV/d;~~/80z(cÙt{4I>6E"Ma 1 pXE$WGz)ڸI:M.'){<*-))f Y+PR "Cхbhlh4&xuz:M*aY C(XRV.*r6΋hy¢W$f2%66#v vtW5^NNWNj!d 4Eނ)Z҆ )|.J!|=+ɻp"|̧n?R=CZɸf|}i)Cn&kgb|LWuӐiJi чQ)g0,LŃ~vѣݫiU;+A:]dݬkUX٬4&> v2Wwp6(c&1 )`rn|%?wvwoӿg}Wo1Q+u Gu `3?{Sͧ&aTMM 楸%~żsWE'ւ(}?tG.EDN[1)tu~kVMBWb}Wl 7e%bz Q Cl'م@ǴYc<#R UQb+J*hr3H9 J)lLX'x>6(%@L8Or'SD! KX;kqTEƬr M&udXTTqGkĦhIscllLj`5 BV)k#/Bvm=]ķOPj><Ĕzvym%L1 Z:%?M.Džu8>hF4- SXMq2<'?O~B4"PT7ZuQmbNk,tDaeGσe889m |gʷ1?7B<~a yzhk s3 ?Gϯ1CGC|Z SZ J""H C$)l4p:y* :lX̨3hQ9b0XJB@bj`PELM#IiYcۉeٛ_?6{dc ^1#1K%ȸuF:Oӻ^}5efW_#k~Lka_ʂwݱ`y,X2 ~{o2_P0'Y9E( Zh$W*xetZWZ-S~Iڥb?&S j?ʰ3}6ﬦ7驻zU'(bB .,REJB3n-KO'amFI5 Cg`x/i`c$8UB( .sO[M#4U yw[rh*567ӸBHEsl &1{ou1ƤT%bc-)• 3ňX0N{Fe`|t3:˓$n(Ac y"&hprQZbDiL%XXק<:,%YT 2-kf'&d}5XסE˘ mY(7B2^an*=n Y_zYRΌ?ߌU`lq뢃"`UxbbJaw7N00/)(?8uJUčRbPׁtvM]]-x6 (of{\$g]W%.9/3I+'#`뿗3QQ{ )tpvAX Fq.m9w2)>Lr` UfWIHou^r!.B_ E-p&(Fg3vF3F 3)1qxU9N7gx8Z| 6鹪޿}uȥT^&pF)W8E9"~|3MXN?Ŵ?}3Oǹ8SR^jT5(袶դ|vk7=&5lʂפx^=IhvfdRӵ4Я4>+/ZP%믇fI=zOR'bv1Tŋ]8F`=ywNa{0“,0M9v$#g pH:L(u:Ƃ]-Oi3kqf;ōH!0ڙd#AhU,NX (("0(b,&]<]<K;$=$-^u5MlaڍNr ~e,n ,.*fԅ'sO'ΐKEr.EВ)E 0;g'!wxlې~i^zvZBiO^t1Йx,ak U cJ#68R!7g)Oc.*5X@U>7zH*WK҇o[|ʠ tܛC4->T[5treW_qy;nrI5=гjc[#fաa~Epe;ʌ(mP+~Q-iT;M:"a(_#kдB2 /TcR)0yc;8u%\ i^c./@aW4Mj}s[x2A0ɵ rO cF@1b"iZ+I^/ժq߁T;k^a-!1KyMZLZxl5ݶv;w$'>z `q{0F bLB/I|" u /?`{k; ^_EKc$V"Xo`8Mq4(wĘ ${JT?L4f޽|'W(ڬtWa+&&RJ^0Č|vQΘq(־]mvS>jrWjnևiYnhoQ=;d;N׆g x[pئ_Tݓ..oQ!>Ξ{pbn7?70^{GExRqT`e݅p- 6lɉ!䘭8!f+21㿳U620"(-:"JH%`*(FDVP0T+6^ ^[q }J!SdI)7lggyI9{Bԛ7I'-rQ6pp*Z6KX3LSesc WH/!d`M blIAu ap) 0i)J;f2Rʃ?{Wƍ ] UWI6J!Q4 L<߯13)K#RHmʒ Fwii5 ,az#gmHH0NƜil6LF#zlqf#hd8O<)JWdoq($B]'9HB$Nj5JụD>WG n\@VârU3FÒ1˘C)U *1E|>ȈR;&xSI* :ia5s:U"\z-k_t첶r`P P8FIZԞQn?tq?XJ*!il0>1׾T=UΎ=spM{@M{9;RxQfZ]\6h7&V2=gǒx*E.Z4}= >QN8e%#O}#E{^>|G[][ }~ rHyf->&ἜZqjm7co0MSW-'0UeQ1>t"XU1cx%wU$WcJs+VPO,:!l"RsԐq.1+ˌ )8#6i) ͩRRV@>t$lI !#zj Mi Σ^}>P 3);j9#%gU>B#R{,˘>H*M+{%sd̽WJ^+R2JVQ+{%sdY+{%sK2B8/hP6_ |A/Qkqb*hJ*hm6_ giQK'%6'%6kէ%r䮃K%6o7.1J}iQc9FiQc9FiQc9F %>s(Q2MHA)G]t.RE ȗa(&C. \ p.@jjN}4<,k=aE£TVS%8QVT>(KaߩǾSfI}8&ݣh"YEx 1jxr H͈NJ$BJY`9 Wx^ڴx=ir7 ɀ;1'BaLW1 @Òg( hԢͶ! M\A`ZEt4eV 0]T2VetF4:?* ZqdQt!<3X%&iO7!R4QNHxK(+"_5f6fIh(4P#f/:VɠcE"!ь&Adַ,HA*]a* i9pB!golLʻ*\>6~~i|(8uсO 10NKˑեN1EDM2{jc?ޅ7, ˹ 9ۡ:P&[Y"wzQMPk/sgĐtrP&w݇ dx甇[.ㅟ\ +T8U{m(@Jz!g|$Dx'\ W 0jfFtɧT^31pi|WGīՊ]%\ٯ֟k7QW=?ZMð]8wM-'`(q=EK+q Ɣ9G9wsx ?a+^]_lgwôܬ{?Y[n7k AVxKOzq'q k$M;G.aX96FA8: W0bgz1}sTFu=ɮQ;jhs_̑lO n X b~є>2D UjbІx!ЅM>JǸGiM,ͤ!oU%ueRw nWNjfcw#T}@hcWTP T%JUx qJxcGlx8k]&p8πsk?\*.IN39DZxѽ4{A+1ɜBfyJ n@K&G# KMO.J9GfљlQ=[ȽU)Ryw`74ߌiѺ@7m[?LaM_DA9PVB\b}B%5(*W&,wTX4“h-P (bs*z˭%h67 4pHQrFB \42W@5*FrFf9()ʝS)"ΒԅQz#gR<h5Q7`V~oh7qqYwQ)ʠ`Ljc>ӷU>a\"eE1(QJ9ϲFKZjIyTNhDa|R Eds݅Iz  kO^3SkJ8OhSXU^h`sY`"H&RłpaE{,0;i!}Eԃ=MH{"{PYۛn?/}J.4ܘ3Or S ADy-&GIɁC5p5,a-ggXL:!} >/E>l&x5Rcrw'dѽ='M׫|TjI ?F' bv1mr$*fGM x\,wUuۅR4ȜvYnxT 5gņC>Ȅ9mG7gGaG~djy䨭MjT[1vszkzOM_պl2W[]ϖvWYկdлg`38wShK%H6[-7u!( Y'2e ÈucЦroƾroB=g ckJ|W瞵]2bN?e"ܿ ]ϵ%9ii[[ 5,x>]ro6cFݽ7^vmw:鮡.y:'uټN<_٨Y=d%aIe(5L#|'x:*7y湧nYMȯU㯗1/$iqxx~PJ2Ǝǧ! -IMsK4gt4  BA&&,[Pʪ)nqUPSRUYفmOX'WIV{v#wO=嚵;⡷*jh$C۰V` #vb7$DMwD?iU3bQA? D%`y2CtAӱ?LgY1GL?mOgq$H6 &adrԳYaĶcDs؂S`zD 6Q Bb%#1S, A94a 5Lgc/gㇴٙlq~gH!`L2#AhU,NX (("0(b,&//LܕvlIlwjzkg'=6iuIp9qOُ75:~=3 54-+Oa3c̱r2H^r-Sw(򥖜N)bxЄ!I8qy MfR(!Zeʹc f$c [cR;ĬW逰>9HysQ bKCk,C}jPzizLi݆_ @A1ҲBP㢸koo/}l>σ%@.~_+j51gиl0He1B-`A-P,5h)͐jO(.äpIQI)%3ZX&zGG u"|aU(+(keu t3]mFZ{BZ1n[z)'$a\5+7ZoSk@Q#[fna*R Οe5()b;m6tDf: '_~|8i)߭388)q0%dX58sk 8N|u] NikÂO8 g71{u|7[ Efe7:-ڊ_5u]Ogfjvh^"o&`8SxǢ)26m^/*$)G'ˣ]=XSS|'5oOn; C,dRbJ)h <*6dtY8nB^ScP'Fh" Kp0QGν`eYHuSF |( 6ūp@s|]sy]/:aL>?!}#\ 9E2hp4FDQkD{* % {Dol׎L#<]o|Xs؍=mp.U/cH@Tn\cBr*ƥ ԔLh]Db k%"9Q4ve=iEjeenšZ gq$waJ#(a Kbg 2W v1CJZ$ :dZ)")$kؙ81bs_ZLi+NeR|jEH򰋭r;]j@ͻ#C=ALI&56*jjtr{oe5(; :͈c~*KK"sG8d?O~" ҈TEOklVE;Dc@ApMH}0bQG"A|0З)Y"b1h#2&"1L qpT]}9$ɷrf[V3,]:_37l=L矞% M*H NHcԕrEEH@H5Dy8Fݾ!th58 QAsglrD`ԀQE.ә8-^xTU$lێٚ́qE,Y*N=BJ@2bw_߿ysT4g^iow_fOQύgϹ5ãU,ׯQ>Xj)q:۰HmA\Tf6wF 9>K45[v1iO|WS6%jԦмmP߼: ./Td˜R "ZclGs-:ߕd@Ha߀ L{ aseFXSjm9LV6 ev/^})Gs#OljZ5ݞӳ!nR:+͡C;6V^3[g9fxZswuv[:K3Ŝed0y;'R [3宆yd?&K%2{n2{nf칙=7f̞s3{nf͌=7f̞s3{nf칙=7{̞s3{nf1f̞s_칙=7f̞s3{~ !ss2df칙-f̞!c Wd܌2{nf(+f̞rjgÞ5mwIZ=7IEfs)Ռ̞c̞s3{nf칙=7f̞s_v G̞˦e̞s3{nf칙=7C @̞gf칙=7f̞s3{nf칙=7G[Gk~Fi{Ł#!֖KW2Ey|&8daEjEhdMIoЖITJ!*Z&4Izrܼ^ toYf}f]>3}Nf{Z/ϚtqFyMr1)"R!R'¿i% ##$]n8;$${p ibTHp(`bQ;Xooմ34$.֛j:OZ;ʹK$es#~jD¬I+B[8G~1xP"_"\rQ"EAfӍF ]Rד=K< hӌ+$S'`i Z·H9s,{NHTF m#T&Ilα@='I `^hF߆#n hӹ' ֥?E:p#H9m%2b gޜc:݄bx7Ô3 ܤ kjqTE>n~~5oJ.UH6)<~5à.evL~kZ:Ai.O Lс~ B-,TF͙= /l5:g2>N -?pE5,{úBU09m8$"^2\c1;&2FO?*\}?O'%k&S3n>o5Z&W6 \2@cQʕNQ "oL)KK7Yڹ8VRNk;t׋" f$b) }`RbY#2?M t/~7OZiXP}.Oo7I=Xϴ?=:~ZҦ5]ߛp,;ix`:oipX)d,,SIrGY η J *h1tz-7J^5t8EW8D_Bct菺vn凴C޳'m1$w&(j5$/-eJ_11c/v,Bݑ\{}p`͛n3Xy4 ޸cZZc jZk(B=~3m[Loܨjxk|YlGnn~o b0gS9" [bVANQRV8 `Z6(OV(G//q2dz#Aoz˃u΢9mJc 9e ܊Gp~٫H#ULXqzI]uۙnE.DsLrX=$6!c\xKh%4nJr `ƞ fhc9A(%<8c_uh RkK&/5ܦJAcsrWDpOU4V#][A1bndKҲ5gYZOxwvb+"Сq>ͅYLuwqÎXCٝQgaj؟C~ qB<=7:se5+YV?\1c=O ~21)Gef44#*Y4Թ1Ec֜l${)-QpPւ孆ܪL;fݧp@tK70|;Yͣ%mx;}Clɂ9-"'^q<]WJ,kOr"c%$9I#o/R֓Tf6.͛$'Tǘ2S 9!'blIbVrEX4 IxjZS)ek R5c[L Bl)c)1,xE#QyZCޕq$2 !X$wml /6>%(RKR !)q(1 X8 JAiQ`%+$M|Ci,B+(8. åJq`N=#dGh l4V[T>Z 7.ɘèJ  H7sd Hg x(2J%1)D c;$c))vkW]HjR}I0[:ÁLK 2 Il!HၱkT V0brȩ'* \n}Q߬j41i@} !FAJͱ;(ֈ( Epiy1` R;ĬW逰U`g)Oc.*5XOn-GZ#g [ iYӒk)<5|"ap$#>ZYe |[˰h4]Vn虬#J[`6 (?ʯgA13!(cc<`0C:<NcʠV&VVz1)[tV_iMLr܁'؂ƌƁZ 1b"iE*\$Mը*\vL6d+@>A6viL=>YmBwoF\_fK.=,}ZlM= Ó%;?)q<_x5imsUVxմ[xkSL=z>VyWnޓkpQZUʮNeLrKM^v'uH@?f#5!['Vz= K/o~܌Z&KmmlFe['arQP&(K:P9Sc9.-wmV|̄/ N}Z5]ۖ{rn'ND~T|yKfr-zÑj#)w1#DS H2ꙅ x$ )[X1c2bphj4BZ"Z>h5AJݼKD}>)94d]rYZbUk |W|䓢tlQm+v^M>mo8+AMoR.2o[ Ɂ brIcЧu^nv=,ټrn]6wz5nyr~ Kʝl[`m5˯踺D7 ([/I0fuuٚEmDŗ=ely]Rqs;)SOS׻$F9yj%7ih 5(O 6eorD,*N9{ϔ;w&x0}\"պS(EP|g=6oq&?W~-;̑hkDnmҋ*Riu>w|Bޡ cOrj"FSbIKGI(B1: ':=Ncq,z̝vX-"ro0CJf/hll5Tzy,@8 pryU[Bstv~dP]%ٗo5dzpQNOF BSɭ[ Y߫ 6 FmJu5lᝪY<"Lkuyɼqv1D<_/A˹z6s`דjɹL'/BM#I8R rHMÐa4ifYށ 0: f0bɇM=٘di8*AG_iwUX9?#Y.ΆeN K:zqzWӻW'?~~?o~{}wӓNN߽vo`&En nL¯7 |v[+| N0OG"S)S61pH~0m'5Ṣ]sM/ aYdZcac yM>%܁+&9.]MvF6~M Wk,/ZpaÅH(=NE^.^gXSyt2u Bj<тJ˚}~ux r\5-NDP1^T瀒7?&rGj,DXliqRjVVV|8MGyj 3."Aelhgw^sԹ n>䕈ݓn<`T 9S:|TpKKoE?;:fo 9y W N'Ng@wod;._.ׄ3Nʾ&]_tW%Y@&|oU"퍺J *Q>E]ḭ)on*^qŜM OGknW}3dK!seKt@֎;;0֖Lkjc |6Wfh&X$R[uP\`$C8rNB08Tq9UMku7%uE !>Jty0zmtT n`]DTۍUb>STaqÎUzveFDa!^{r,.5Mi=E@}vX 3%]5"w9FG =^+k")c4׀e[\?*u/;HT*Ҟi (h@='ILEM x#?`OҪLӱ,ᐤ6ev'}Cg1J&й;Hf<yw$a3UBuw-hoa:Jl3èr{$3=r&/Lv A"&"]o@!· >$LS% x b8(S HVS 88%pSFCT(%Uŀ%(g2D'HkKI DpC@,s$k`kqz{kQs紤`ɗ]VvCwVhz3Eu5|jxc[5e!k|QCDXI{@ll@q$r:ƅ{o8ݣXQYSP8Jy\)Vrɾ }k [̭[&W|Wi_5V83/<܈Ҿf6Zpju%&: #[hWyaiK7KmH0'ʒJ!qQ !: 2"Ѡ1roL)({C1*!jc NPr8Lj-$!X(OBr`9X>2UkΆa.V}ڲKHƃ7!pTEbYL2 )PNͧeӚ:"U(XB HTX?K$Ŋ UoEF50D s<Pzo16rn:U6,>OwFR !Ȣ=mM+ #(.'&Iua欷 9eø/|li;bq[ǃEV,"d*eA"03Hې@ SF!;QZ 7Rp JS`)7"xL AS~G[bYo.Nf]ilkiaHv`Wr# -.(-|>sĴy묀r>x//&#&52o!']ml U=?i^ݛAؤ.ZEWݬ'5t"< 1hNU}'i3Q uHi&7xDtYLULtNy8mvW0iuBޫ۠i5t97_#G'G qԅ=̘1K[\T*E0sn)rnvZ}|9z5ArDPORA BN+uk2H~H\_uCu@f[2q]v]L^Tui*֩Ww}"*R8wL)#׺Z!seֈxXTj9s%c{6ɽ{g蝡*~ΖilE5w1GO: $dsjTI7(ͥF(W"Uҫ/AHDsA[ i넥9u 1C$͑#!".Ts6H֡Ѱsvp*SۓιmܒRah|Js,eoI.Dr8mU < j7@r5Hcx`#q@-r`^|tQ]!y. R%M!HB 8QȤ]Ж`I(FXs6ܖyA?hf2&jDe0pf#g4 S'֩RFSl$/eun&F#MI)17#LK@p  @58+*~rkj\~u nJcBJLb?P颎(0M%d f#9"k [`;+6ہO(_Y̳v! j&VGFgT}nlѱG^.^+L^_%^[`Ey_ Mer5k_p|GϝIkg!WJ!Q)E2E8,?rѣ 7fŧfI]df/΂H %<AXbc]l*Jb LCCC N˞M<2'd5Ʉ, _5f'=gm9odLDZ[\ET)tK+7#wg"xC0, i`V؃u<opxm6}.{p앹#)sN # kc4_D09Oy%3 ɒD2䛢LM'eQ*MSzJ[HNgQK`6gEwud[;8Y9ŗO_ CGMv-jIsuR~N*7wupo.j mn j(jDWbO{}5w.P)О)`)hVH(/4F6(<Š^h[Mfx:%qukq͘%f4y/SmБQV=9Fb(LrBO\:;Uu(,h/bюzO}O't?O~"KtJ49)Q.Bt!CTOQBA_IֶL@YR&(1{[$)mR IzDNV`gIy//Cw73$4Ǜt'+k3`xHvuw/z ]:l{`Yg ǫG«ZyÆ . ώ@7ҀA"yhmcq羳'iNwk9H!)o.^c\/5 p%9Q8 DYiOuړnwk5|<J:h=h[j)Z]G^M:W&Qd(_Z8m̊[C@8Lzeզv$:8bt=GźI@ Qib7Y *xS |PEj!"c~wr\gsq1Lu}% IU x#3:IقXQ_H.uQ{C,!'G THEs26BYdNZȔL.-0@e̪g^-Y@Z˫411(> >-Z`%fIQ S-T&Uxɿ?<55̞6/ۓy'·kQSWϸDVQF()Qg7ܵY5]ݟx>W>V,Y%z! X2. yewK̻yB.)e*EyxW/0K Kkz)nBLggϳ) nahxtMtw*M_5aa>g>+.ɷPp4-=55Sc.o_g+:Yuz9,W~1KOXa8ܫ<3%'׶fuKQjm33/ĔvXO~6EחwqmJlnu6u+JVՒgk6lX|:c޿ʠZ39MLO1TMTjY&9r8y<7,?_~?~ǟO? XuAWwE4?ڽijۛ5viv9Ю]yM[bk8?[O~?LO(yߏo*@{Q/['k0L4wcr<*^QP_*^v A/~Pkv16Hij$ް+ ޕ.)  u .lyUVRN@hJ'0vvFvC"ZJ:DI"AXRrWjNX)/% 1WUɽ8}yy3x`k;[﨟LAz»>S?Uq\wRg(>qW_oM;MTLju|?IEx? _Oh)k:~+x< fgAc֔L\4CC 5([7 rQ`EgDӤS6'?/5!wJ<\}3gtqm}U5ݬϛ; ^"dQ7e C@=%4XW$rF0VTBpJKV.:~)׍9\7pr}|CsW5)ywM6QmUs\nsx~5_Nnڸ&jPkp|X*ۜK㯥It%<ٞ~l>k?ݫ4Oj^^4gJnhC};gAYC;b2薊!(K1gEw Y]dQW_ί?}3}u v]zH/TV!"N͙F!F~%_sbbbf492"A@`Jx&`1h%|qX6UJ15!jr7⾀XYXF5ɳ*zQ+ugR6q h/ TY;o,7; p}q`?Qo䌲ϑb:>Sپ{ƶQZ7j!"?#F|ь~ŸD2 Ai rTM&|t! :DQJT1˘nٴ} 'Y2IfIImIS&Qh9Y]:ug'ɾ|Jп )ż4GK7R5OW'g*^.,t[H2 ~wA-tͻ4mPzhH(ZXdRɤq]Ӂt/7A&e fB %+)-I0F DD=ANw9Oӂ8{>l>ZO,1ZŲ(eYK@+UIN1 e(ԅ-hOzlR9B#mTZhMJerTB"dZ"H(Ia,H+eJ(XH %IdŊBv!$?pJ*v9J!Ȭ KOD2^'-dI&V`u~USՖV-ޕ$ٿRlA=ƌ=FX`{ !O+[modUxHdRAUʊ|"22"*xaI?cJ`ٱ{i%V3N+0S`!'7B J1u'f4 ׶w{2^.sӿJ`hVtb 6%K +e\:R!Wᶸ14=zb Ngn!d t[N4ήњ0JUb陚EjSvХKOb|Op1]+++;VFߧk*?={~Uq&4gaTZ|p_s>]q9UmjQx7};s\NKVie>k?!-/+ 6 FwTT#BLϥr׾b7".iQ~l62x_@.l' f 9 rG(eyE臭&Q j QRNzjj~N' m|kP>:0쩾ks<אM'| 5{tjv596ηzг˴ y*Φҥ9wlQyY}Ʉ1%`h(B e-a8?Lx҆R{&<<k*V$X;r ,%Y&R S41H8Ha~_x htiseKdwk)X`E^!OE& wY4UTsc%bDO`E. Q0'fOؿ޾KZaKJc&a\T DAtgRi%qf/}թ * C0$,sFhto[OXpu'Ϻ)s 0 U0-p?{N>T}u5c;͙U /jz*wƓaTonP[Y&um h{w%hm,nLT )^6 y=.]B779o&,.u%|ЌhZ:q(J`cZ1rHeZLI|9E cDJP5N5N:hwZd7>IiOf<89|(/./1L`h !?v MU=q4ef!?SXŧ.)1CGC`js'G1 J""H C$i4pth5pX̨3hQ9b0XJB@bj`PELM#x%qJKׅ]νfZmS)r#a3[Oקo5"wvO<3bmt%\ Q:O7GjbkGȎGnGc#`=?MH/rEjP0'Y9E(-e+R:|+j^ 90L0[vs5)Vu;M:Yf [Y k^O+Pp'(bB .,REJB3n-KO"߄i% -".ߍi1ދ`um?(8qpJqS$!*`bQ{&CD{fhpn;Ŝ(arWX&»-FW~nOb[S9a^nDWs yʴ[xO_ &W:p~1i_A|Xr1FJO Ǘ"Ϸ< hӌ+$Sp4O-!qk' $n\/@<ZgF?RK() +)#K#,zU8nLK=oG}QbfWuZ9C"Cۄz)muvE.d&nݿh**("R1U6"M%V sbM5*D 7a|DRr*%._O;8EfW8T?}8oD*ءwsmn~)Kt#M }B8^D/tË~^\fA߇a?觍%ܸ9{S]ǯixt¸K"n4ԽY ;!t_??>uqf~9LJw{ϡ0 _??.QO&]_~]|q%1/dqQʕ8" t}S :Ѭ9/5~N=`Y{8QR^m{Td)/uPRlh|)x˿o:MՓܭ3=҅:^ϕDq2^^u~*upg?:h&1+tNJgCxa_xB ǨiX?%>۠s: 3 `D;L:{:7\N+?.i-l+ڜ/.mxw4}_l>_o8D_}Bctk٤hl.9_~6=6ߝRr#hy>Z]4ŮNCYsBW#&EW!^+54 7Xvhn}aY4q훿kJSmI{DGPм1aC}7HԘ _G[pO:Uۓ@MJiiڿ5Դڮ޳[.mos4j5QD 7Vxu,8 9<$hWPsә3v/N7O<.%i`sH r2lI,E#T96(j=9Y<ۋ<ʴ!<ǝYT0gq 9-vNe#/@4Rńxۂ^KEyTȳƎG(H=V-Ji}Ryv\S,4҇ʞ^)^eB׹Fe] [ @!ҕD{]2I %ZCԷ[[99xRx$Rx9.q(:XCTU)S*(v'CJ$&x$ )[X1"﵌FMFS>FDf›Mn<ɜNz?e3NyOa(R0:HG6"a{qˌNv&TrLցXڥZ*>s\UU 80A" T)JfnCңԺݣ]*l2ܥ[A]J%KP0vY4a#"|rVk/;^sXer&eZΗD*r5 (6 iPrԴ_]S *uR߆Np{wz_wK?_Crrѳ%{$X0gpbBGep36SNV5j4ZM-z&J8B(ˋ0. @Q^A6ciD{/gҽ`YHU_vOG@ׅ4Ueu+1}Tkwnйn-Fk~Q%MTVPV>)m%!a40u~/D%VD(tQItR*+QISF%SpUDBq39*C,._(n7>4Ζ/PVYL.|{U0A:YAv*r" ڴ]֕1n^.ybս|@vtztIe=nǮ*C!m2y\ syj*\ [\:E$QW''`.ٯW?_44ҍ]A'cAcbUUk(T\AhUPDrQBNv2AJ6;c.))t{g]m+C* ȒʋdFlN:_jң%o'z^MD~x֏9 gt:ZtǓLu\7]pdR]QO.B771iMNG F})3Xb-Z <kZz,*͸c1N[XL[Hl!xGc6=KǏ_7Ns T](|Q|,mrcO _tTŇ7uQЄ`{.ӏ1y,ܠK3hܸyh֍VV'1S{r],aBKu:_[(VLcv])2i )ZGi/se9M^>_mrsYVD S63L+9.~ي(ޛFfprMٻ/+8W Ec /]6?A|Nr.bW]XxvkؼJWWM^P?iHkui{!j˽saN&?6RMtFYH5Dhz~iBuvzn?P@#NyS^Oo^n5L6"Hg*U8EqQ!.ŭ~Nh>~Y"=Le(jF$'Ӌ4wT.\"ݣi2؎Yzt-_ R͇H׼~;4>E3~m}YNnc1ٹ[./r+7΍n 7݀K?4qV~ݝMlm,ZFn`.]3I׮EUP&DLEXR-:kNee3&Z\جVY9f+,U>ۤXfF?t\J#WIrIZ>u_=X+-N`c!L!vsRYM]ƾŽ߉jtڋ+"%}[B 6ۀvOU@9Zo,]$DdrqOnGHR }wqc#&TXG]sv yVb-tZL>vߑe~zϺ^_>&Tr8Zcv 61=eFb6\\z;t\5{̈ĕ F X2j?t\J+F\ УK;z%@d++E.bzzJy`jW֘Uq0x"; v]\r}UJqu2 t.,'@BՉcR}߳V{ lրh%z{:|wS_Mo,k$r6;Z$~{5qɫpp>p CΧ2Ѥp&ke-~k^ SҷŴfE~C~Mbۿ}x-:/'_KYh )0*ĪPz3] Jdڰ`ٌH1ІUJC m.#\=\ 6\Z5ynV9vF\=P`Fb>ynD6XS^=X8qpMBX^`\~jJ;%W0jSEFbϰF{Wbց:X#W+lW$I Hm;@q%A gG\ W$X Xsj)+VԈĕWA9EW$X|+D.bFW҉W+-j; H.lY-UW+#9M`|pre6AV?dz|3x2Z ofN=8^Gksֻ\H"2f{P0Tcvܒ6jB?NpHMZP10Z6ċ`dQP,`岙Y`rY`vkŶC-W$aprU6XD(0q'8E|,fކGWnՋhzy,z9`?vϋz-Tp%{J|vL?\qc w'PO*7~45@e'86$jNq2O;8KQ:HZzF*Hp"5Xɵ>\Z+숫(1#\`zWV 5t\JW+Q8H0|prWT1:H\imP+oSlX|t*kճhRe+lm6b>芓:X%W+#L< %Nj^lkP9-]# D6fr4Urfڏ>@J;ϔ5ټa J WF\"O|bg+XC ї+hz<[rzy?{TJձTЦpզ`DF"ʈlpEr5\pEj;TZ1q%~&; ˕DWq*jճ *#\`2\\՚U:7qů4d+0z Hn#.qu2h~,W$M6iUZ WR1:H\V2ձ'Ÿ;XIMjahZQ *~xAcM |k;ܘg#<;Kn>R5>dF}~Lyc6^F;v6^lv9dݻhjK鳙Z`Z2jŐJL!N-8~0LglŰτJ#W^)/sZ畖e+; Wn+gțW/W'`/R{KW?>uzurhbT\Wpd+W,wS8t\J#WR9EWSj]\rU0t\J=qpłW,\pEj|t*qu2+l }ORb~qzNm3򂢧?>va)TTi,>cQlſ>IS.lu *Y)2(%tڸ&B@Erfz$o ?JNgԹ_cf_ *jzU^Tˍ\D!<Ԧ*[YQQV֡Q)5QR(wȖ6zJeJ.|iBD|چdPHVj+*?!ہ[W٠%R!5ՕkU5XUʦZT“ͥ_YKb9 AZ`K*UU\VyB:ev֦R"NҁO"+aYE2H*Fm XW2x ^ƖҬUKU{T}9N?]PHΗZXVU։$ƒ"): I{jZHPMXKibZGn":*(u,CV﶐G9JT)2T})Ke7 ][~@%IVB eY0Z~"?"ʘԪ4mPP/K5s@lH ˆ« }~Kחg!EκAFT:k_DI_{&!}NOʆd͵[U+RY+#C@"k1CBcNm_Kn<tMI|'@HK^q֩k)c Quj= -^R0e$QbI$Z RJJDL)+KA@E֊pb)Q R*)jaKtPwmZk^{H[ MHp1O[M"kX,"}Z|3g̚R`ƨ%gEƲ7XeM(TGlO6W*KIn L5RP/q*}z26 P[ nEE F'+9+0N=8 /5p;~I 5 <&B`yUɖޔ1h4y27X"Rc lN@\QQѠ3.4hd ͢3g*){Sm.3*p ar߿`oҸ9kQ Ki)PV}JC.0(]-+-wcN-)7*ZtUTiX4dDfT@6$͐ Ek=>id3X{ eR(U(N` , _tkzUEJA  hW$1vlŠ6AwSA 0`Ɛ)eo%0cA$ d&X4% HLPV"˄:}n*Atoo95jT4O>TK1UM5. 쌌7ź(K5T&/ vR A6gT4g"!(Q(ʨq{A*83jYɖF} * yjSV&b ]5=*)Eh LcSWki ^HHH( 6jer $zu(*(_CXkuڄ t=i/*MVLKQD3f' cDeS; ')! /P)<\,Ч5x~ßZp %cc;-t9/m@-$t>Fju`8 >n9@C>+&9Pmi>jZ.]Us-EMp%eF :!/5* <@H"5T^1 6>)S+W?XG@/> ͗Te2jnu ›Aq*qt/,TɂNU~T}%*y*j m+d@l$SU>}p_̧SqQ'sOYs`I%> ^{@#Bzv)]L߃bTmB%\ ZB `bֶ5P P~3 l`JQ%\PD;ܱ -PC:jˢ9k4Q8F4@؉li(5&?z+Rp 2ԿQn%,BT5y+5U&!*P/mMumgNO֫5g}i~7p߹ֲXlAc2.!`Ǹp?+l\zƥu=Y'DpvsZ׫cmC+DI!\YNa KM*0rt]Į;LN>z/ǗrQ_7ߡAAMƽn#6o?⾉,L#]Jq0t,0h=j0|`Y>"r6sYYIw>WHwd>Gxqci=~iX)Ӳ0)73""GxZWy9Yqv~6\۸7#;>#[ԪU6,(QUkv:[|<ˋmӼx x_ʏe+̻)6wuyJ} QNA7Ơl M Z4biC?4&1ţ@[:P(ol=*cΊ /ρmִچ QqHZ!b]? @_m'HPe̐Wj<ȈcZVdz,XzgB{xjN-kMMZww,nj%ԆR:Ku3wįǛ.v۲ڀ U SFVqR(հ>ᯗ# /#~έžcqzg-|˛7p/r>Jo5ۑ^\{"߶_V~tM0?=9|(lJgOI[hD%524ל6Im_CaUTT4v9&s/zM=U, g :)RX4s̀&ZBE?V]gA^^SOS'P;(滘ߢlDȓyJ F=dST 6m)#ϯD?98ID?0g2kh Q)PClHz'deT9aS#¹~E!Or}r~qv ^쀏3Ky}[ }tw rv on雫oZTOI HWMlE hIX!|ѕ aZ'ϋ !Jer1Tej>+j)uLbGsg}b>_= 潣_[=MRw VO,J,WagUқt$xfS*A˃!C\b.M:+Hw:?{FeO)`Icw۝iLwc?ֲv }.J2eI}T,u916-{w=FzCrx8lְ;6]uףYqIdx$'dck4{V*JyDAHُt/c8gfhf=Ԋ|i``'1g N𠜉18*<11$ε!F =J@E'U&!A!rk0t@ u.F7 KF ,`4$ peXҌʐ %ES\$ xaYJw2pU\õ؁%6Wiޭ2O0vA^ 0AtSy2]walp'!H FW٨5%'V{F*j}.SsHQ$T ̧hfbs5zZ]G жF=VvX0#Y/mUj60$匲>5^;%D[&GIGEw0xW,az9>{O>/h0; R_tp!=Nfv\Gj.o0vnҿt?V&ҹL' t0SeOu"{)*g~<=Qn2EGQԩJz?~ɒ0' 03t0m`r0O c/ KQoTɑ\&2)w?a4.>Ca ?^^] bݻޗ@B1E+W9cBsPH؄߱Wӌk~|h oͭ~bYk^hEۜ+ٻD;y a3&,wO F3Ğ_>鷑K䧱uC7 h:Ӎ|j"N0(NӳQ\joٿ;o=W9Qۦva͐xC $'"=*PpTr*Jxl/du'KojS\ ?DҰV~s4vWC2m)blSQoz_' p?|)wrP:jt6A o$?H7 ڠV2)|G\;y.qJh^e]71y.v RS7;i<[hCrfn,(gY442nR[ ) 3Cx~&v5wCZIRz{Xl4R2 ظ=)7-x&RjKm7e>-a5ae֘Ud>t7ѶSvfe 7 I7 jf7.eL6:LFm"<ۋ<*V ;7[3C:%E2щBQ"zEU`j$eBg:DxF(]2-0joty1۸5eNv2_vo/L컢Z=Vly^ߟOv\Ўkijj~U.6[ݰq!|JduDP1m%e *I/,N^Kvl[%?^{)6ʀͅDYR)&4c1ƔF H˃1IEٰ; 9br^[>罥l!Dy%,wVzcL&ŚÍ}Ls|?XM%>)5#GW(:XvF.̛J*=/};2ӚzuTD[SlgYr,i}%8P NAkML(ŚQ,*X5y=MscQ䵫|y2v9lm^aE+Y"ڃUn]`klakC-[\VI&a:z2r XudpA cޕ1Ȧ(C`:&shRNCTTqG.!HFbَJ1,,fb)ƒG,c Wln֫ɋ>Bo Tn8_7*qGk)0µJ)XTY b>0cTѣn'ݐŞ32lr^:!oBHLXVO?Ts#<.f-ڶCnxk0ըÐhy hG<TnJ(F6y#1Hs[֩ٿyo$R*:pЈ5HP #b^p>$E5hW: wlZ/,20 "= )u "yɢqGPz%NmmCGP􃢈jc#QKg|P`B@%̈́NS nB'#bpq?>ÏfUƥai&<.:D\Zo Fڋ1$GA-vJWE5oyOD5N%bmԒ﫼Wj)` s,Nvn>wqCңLjl F%s`֠}%ȡO~*`IsJ>K,hP୍c!TG`VZ?6n\?wop?] c=~x ǡNʅJzԛcsLZgxPSJ۲Yӎӎ)À?5$=X꓉?Ga/ͫrD' A?eJa8-MRr>$r#eZu5T-;J[u䈃ä;#fiҮ>*xvOwԆ VFz!RVgB8*R%V̼nrtY1oX~Wz \"8."B8("ODUP%Fu2-X"XO׏IϠO>5?O2^|uX&n[]a Q=)N/żI,RJhtxDqd>ǽ!ı-BcZث-!$:j}R&&@4Jzy  15pDe>.,rqa.]5 }ik{ޡ[vM{%]ԼRr=L -n?*l~\ܳ|Kxq`Y򭳰f(z>|Юg_>-Ԕ/Vn6!m6sbU=d ֨UDoh%=EZlّ)[uDVR/'y*B lH)ϝ1[ r<Ì\j\qR2b;df^rŽҜ'.Ec9ͦ'9E#DF b!G$4 fjؿhܒAo%4;?0ݸÏeջQȍ#"HJp&(WZRgA颾B\&pd$r ;\H}ԖhI,XsvOA:O9O4UJiM&I[(bs3 x<(^ZOA_P\PӅIT6>P\Tg=:DŽHp9!$#[Ձ{Um(4V)Tb|JQ*x7Q'-fN@% M!ce~߶}}lKul+m _h:PkS^@ۻ*i8PIy6/k'ATYgL%*FMf.ghp&.w0u߬|J8 4!+Ak+iAzT.(CJ})/ܜ. S̊(A6,.^yglNvJ/g,>4^_{м4D+kǎ;2Fjqxv4̳*8KQ1 Z:1;*WR(S(z.JgD: jk&"%VH>CM${Ƒ%B% &N]>N$ϻ?[A[6LCY:N~ZؚP!Cm7s䖼aBK0\ֵ6`ʂBb-)҉9%^k3ny|-WR7ұA ,k[XWGiq_8[-RO/W%ǧkSq3*0֤vTh`UuA[TP*I֦yq>~77\:Mǒj źimmK#6 _VvAFrϱ5 Bp!t FѺ cen j7[@Jp}b ưf7֭._CSZ:DhU˶**VZMi0rmWmnx Uu]kuq1R)-:PPK|cUVhrCoCt"cehP}*M|Ѷm+DMmZסi ]ygpc{}L.X4t7Hѓ^ 1@דs瑀|rniL2]UEeQ5ߜ䟋?>jXiOahW?I % )xUPUcKUGSP'EoJShO:JyӮ2Qm⩋UE]:K:CcyTh9'"—B,:y^g/:r1P߱Չ{c j둆9R%-sS|)êpD}FYm_ _cNٸ6aZZ_Srh3L탏e 2Y⡒YpIq4Oʂ0W>6JI. %[+K=%ڦ5jM j(,4ExVR(ZsnI=~/Uj/y`Y?x=okoͷ{5\)ޛcWQ̂&tE5E[6"DxuN~Hc['}y,]}va]?GaR}ovT?ڡ6Kt:Rf)7K=xY{ztKRtJ!}:uo^+.ɧ׿~d._sL|Q'_%.|ݭjK4}I$(\o8);ϿwK?|>c'j>F=o<[6.sO Xrթt=E=j˞ h}jl&v}uwÜ+=l-X>m'`g'E&b?ˇN6sCMA$^3\c^5_/WZc>ijjiˆE3۪z~>AP. G)FzƔhr6P Tx+G};iK]WL7j427g銀VbtŸRtŴ )d]PWV[ ]pP(FWk]1-&+F+C0J81"ܨL32O]WLiuAE-3A+btŸhA'*)מiՈt]{<[ߌb=?NɋEk݂ )؉r'Μ(s7IbhMO g;nTTT-[xϳw&i-^c\3#J1uM3cvct+ FWZ֩rmQxtc]0j9,`ZƬ/FWaˢ3erdU?\3C?ZgU?JB]=%g ]10*1"\]1m)F+0 HWlлڗ*)bZ|tŔ!j|1 `qc+ ƥ+38F]Yk+^uQi!)-d]PW^+ (EWDkM]WL 6jNQ {btŸi]WDɺ3F'+dz+b 8lZO_>.!]lYfTAvLYpiz~Z L朷Z[ݡ)EC}[jͼfڹD񆯱t6J]uN89>K2qɪ,OB^/TMwϰNWe9m5bpvZklvqyKY=.e⟦FWZVmӍڡ'ڧ9r;pFgJojj)@?Ţn)A,ںRN%k%Kn9kU[Kqzrq>q[ʮmlt6ڶS[)Dzq:(+ eh_(`I99#33-&?qCT桅-D!0=]+?( Y%HW+5b&nabʼ UܲLg# q]õ/G{YMGҋ=tZ: ]j p RtEc"JkCuzAhbtŸ^Qg]EW&(HW ވF-EWDkO]WL 9,*Z8 2.)"ZSu5B]9% 3C1b\/fiZwՋr=GWxt$ ' O͔>_~m AҐ gBqQLhQ-͔9I rtEN[)"_AqmQxtJ#J*8 tŸC')E>՘23.vnY8S8zd?\;pgPQr*c]aS^ tS@qiK]WLmuZtENZ)bZSSbƨ+mВt(3HQ;)bZSSu5B]YO]L_W twtRhAuŔk ƣ+$ 91bZ2d]QW3f8]1."Zojg2d]GWޫT-NX T Z=Mbʘ1;P l>Z1A 쬫t 'ƞh;s&zgj&UC%a݊N2A%۲۱Q#KM6Y$_*?; ız%oGډeW>v^]ؔ!}ЬF+s}_iJ@bYgy8ʓ%ʐb2I(mD=A[is,PM M5C<|iJIec@~:<1!4v0WsFOaQ;~ߞd))M'@OxӡOަ&nonhWT?Fy$ž{ݻw 2 '&}Oh` s:&{C kᡭu"͹)gS򲤅h&Q8+Iy?ֵ̓u]tMGӰHKJj$^%ZqӆE %AgS-(kOhY=QExLQ4vÌb_UXAbܨȫ^% 2XV`MU@lDFy. KB`(2.12Esro/܇n9"=q.zHXG"\JlBq%&zqyNJki\hҬƺ6]83^o 1{4nt :'My:We 0p`ҳɵzkCy5_Sݫ v4#QRf/Z۩t}fT-òV¤(xVU28E0c(qE!2ϬV D)pVt+LP à cf2bFsב `JTTD2'#ˀNwZI0A"A$ rqcP Nɨ KXC MKI0(#QpZ󀑍FOh`B&:#E YV?'ۿt#A i"B.(= ;`(AW 82qP"*0 H R(߯͜C$ &xi,oi:x.L01书"^HK> y!p^?EYWK+dqZd ci0 C38b8B *; XB ${4AJZާLIC?z${{r0Ok`MދKs5kêA;cH3` QT J1u33;O@ u@糴_%P0 8+:mqkǥCR2`. ),2$qQ('J8^/&}>gCS+Y/A>Wk^;9vW!d ;;Fy&$MqݴQ0SH,TVZf&{vnvSޤ\';z8+&W5KVumnmMHx]8(e$Ú}~F*t +fdVncP- 7MaE$gW'?=/^|W':Cg ޙFѯ7жUSnRhճA-PRmsk-@gN?=+b*x:ܬoLrH+g~ b~^u"j~@BL/rgV! 024V}WHhb'; ߫_NRA r2Q,:# D-%DJR*vҷv8U^^KϟtE~Il&&ת+ޙ rpuʲ4J`  Zd+$a#p.@:SuZYtxG#E6﮾xb׀niX]01ir]VwZY^&׫:Ӯ-* h9nkFX(J3h⌲ʁBfnlgu:Õ`M΢zT2F$2/= %mDv{Z2څ'VUSƩ`Z_ϫۯz'toQ bUlsuKIVrm KlB*'5sh#W-(;7_*3F>yvA˕(7aƽ'q5|+|$xEbϦe=uġ,J#kU>oσ=o1z+e^$%XPvQXJ<)@-L )#"hb F%pHފsw]{.]GSL֌mCy Axi3>3ں 5ٻab*}T،]MGa2䮝w.ݹ!V~|sGH|5fp,c0"GM,0jnDh:vF6x&yt.0gӀp L1a|L]ٚ5)\;id5j]c[A11T)b+t@Hm7g)Oc.*5X@Uvg&Tnr.LV"Zې 4c?)#>>N \kAxr7c|_V-ND+f ¹]@/k܌xa]Z'z Ea-#An|`pG0Ao `hn)1H4us7m*Ӯ;v8͕qh N:D`{:()ݽXH>&L.DpCLDn,,xGI4 GZ^So=CG~풼uUy#@B8U;W:/kdP͉˿[>]4Hja$WsNQ9*Q@bўJBDaGͮCmCf:7yVN&o1CGC|ra:])\`[ d:< vWe)]kxM2XkI ׵$*W<\׹EIxW34t4i n/0#!K3Ey|&^{y?\~'>U#̉FDN4 ~LrWVJQqEK!<'iY|'}1(~cdj}Q;WO[Zz3oXe`z;FyMr1)"R!R'Cn0M (4½h9(bv)l"X 6A)8qpJpS !*`bQxشʦz{5e&7uy>$7q \Z; >lRYfE$̚H"Ձsv<(/t:FJyP<(j4٪z{< hӌ+$S'`i ZʇH9syZ< 7H|zἘVArOTЊ>xRZbDiL%XXOytYJaѳqeZ*؝ݢͶ{#C-2{SpH`:vEm[v@Ș"[Km($L|\a@GI(mDJPX[is,PωERj2ؠW!oу]ٶi{pɲS y@tuL'^.ʪaW3읙:J980b4,ʓ1tI1a;xuxнPm5:YnX>zeU nRKN6~_ BrxIAUN7\"$aNjdH;/kb!XvXcgWǐ;;J^Ym."@yQW{1\]Dv2>{4Kc&p΀r`. 8D t=>RQuhoKқ7/.}=ts!E)uggtbv%.Mi=a&|m؛E.[.=ty_>5'_j܁SN5[M97v4{`4f:BpFL')qz{ywNaa`$7“,h󬹭Vmn)+5^ .jugt5ϢRuK}A ʀg{9W! x~U?XF|q @pƢՒS=CJԃQMi; H4gjzU]'9i'j96|!%g.fy3͙Jm/ZB{5cXEW~K kcf~~tzoD# #\̢ 4I1]Io9ų0aWC}7P0iWFo: ̷\=6fSL+M8,{Ũ}rb4&&+eeޱ(&l%a{gN]w͊àQC2x:2*N6/A/n AҮdq9 fXu0U)EYF.x˹wEe}(ӇPa[Dv ".k=n~|G!%գ/% My/Sxc3^"6PfTTr\#%<ДFIg@5ZP`'!wx\3.&$/${OQL[<ȡ[TX{i-瑩c<ˀS %l)N7 6Ū-t^4Oۀ҇o/gxW1>]SV!h:]Ѹ, [u<6Mru{7=8~rjr</(ȱaˑ=ei\ݢ[Zd݋ZXF[@R:@7Tk[ԩ!4'J*b"\N!Dr\Fi$8%Cn8$KއN&lsKzo;TnEnMQ"ҁiZxHλlUيZJ.39KW7zhjhjէDN_TϒtL^{f2F AY(3$ZrxrNZr(S.{+TUt=0{>l ٮʀ BOg'4`&ӫx@p6J1hZh4j }F/2^d A [ZxLI!Z<Q[ $_7`7&gRZͳYw-WZV֚Kqq>]jqVk:B 8:eb>-ˡELc! \s> ( 2 '"sM,D`%}kˍH_BGM6%Dd5[ (G\R Is栲LV9uWoixaZRbפW$;{}~2p,mPﭠb`nޢM[& 3B3,7еdZ(GݐGБLSQz /l2%ku=×+?"j9ӇvF^F%Y$-eI܅VH vG-}w lDZTD Li G瞒hpnmM|tn' vinKw_ww3*wyZusu=;n0FupIh]qt3ݒz9lr_.b?] & jrl! tCՊe{Q׌(l/ZYh`9ȜY!LO ,H#XV%%GfT׮Vܧ Ks_7fٶX,[.Xwt;"76ƏxUk.}۶7|* ;K.M7b:(Ge7M+ Ѳuy啑a6n ߁wxO<r재IJ؋56{묿!Fzn}k 7r%Zh<=-hv 4tu&]GCɺOt'ߥ f'n}` $o~w /3tCO?;屏;KZ*<LOJ?2jh:#}Ι!:/A}#Ǻr 7Fr~ ]?Ys/$->V; 3mVԞϯf7'9SQR6&D(q&,YO҄!P1^1^wM#ϦErdJ^cڶő)6g-Q+@[d/\^mCa;t} IylTځ`̵["gN/)k>zW˸J1 JFR;pJſF^FGrb>}3@ >ʈarإv)]arإv)]arإ^>L- Zen+"w.ucSdɅRjuSjuU4CJ(J> QyX09K`39!V,+0\0kE4zit|l'k^˽+kZt @$].HʚFMV]¬ ΒZuXenE.K0Єb9S:Ev5%9zEW e veqԧut@L238Y%ͅib>G3eNCoCRmJ<`Nh&)#UV kcr@lɰ A Һv2!3gɄl*as @gX'd_Zh-jgHdM^~ˠ" Kw0wհ<\2_jyK/|H8Ȋ2}:斌a3Z}%;>5ZJ.㱟"oW/qŹ_zEUqP:a&2cIr1%7Є^ ᤭7ǹvv~z∑MS:7K*,̾_o7ww^šF*FB\qQtzi8%gTWK>Lѣnxq[7՟3?'lqUp1<_<>?=V]qMt}6!On'wuw2|<}uo=yUf1wmI]/!x76{} \.U"eW=$fR/TK$ G֐353UO?Ouwpb=H>>/}M{EpyIE +rơ|B{%xo4KctR'f'g{vB9z}˟o;o8#.Go~Fo LS(]V&i'ť~/-֥zKʇ\Z׼u׼qup9KO ɬtkg=dը9Ie c/.BfSΚ'0?EknYL+il!CGVu{]fV4{ƸS#dI|,Ndf" A(g!8X҄~[/&(=餗N8, imL:ouW}碏,oĐ@ti*3Gi96ݼNyI=}W6O}lO5gQu5禞2uȺX7iUtgCKwzYeVn(sLl5룬E7ӚqV-k*њUvY~FY9Z.4eC3}VcO*c4M*6R;KMFTYMVo@i})퐹niK`.ӓy8~^Jm7cBC/ Z=].!˅NK *˔aL.]BL8)"#ӁsٕGD*_ߧh ןzt`ɵRs_Zћqh7ӛ'ޚh 4sKLhKg q V9o}JB(XÏ#Ɏ{g`oQw.㲡sIyRgL 37iN8WP2.x˴ 6h0(n1y)g:![/ b بBfDpgc@d?cާQQYdIġVY Cܦcp\Drħg ]y9"wBMD]w8XtOU\Bע#bԉins(/\EɅt@M+ʢb) % mj39xHS %^Eo P9Psq[hi7?tm32^O֜iJj̴u 킸KZȏON SΏsMdKL&*߿. J˗aX'K̥@JVzEπ@(̭HX5cr4 ).zR(L RsΤǹ.HFj<5s5R yƺX;,|V,0&Ɠϓ0LV?7.\ˆ3H>Xs 9XRYJp"fZ %QPDCU,1"Q` Zz-6L" i:Y0Xp-sGl?ks[Pwڶ2j;{[}ܐ#b #(EZ0 Dblt"e `lNUg\$L̐KR!e!HKW|Yd̹[Q_vి-X?DD"|"vㅵ+Q0[c$V=׌3'1@Ey$F6D{dCl'ƴ@DH ʤ@JZHpI %CT͜$ѫf\۔%"FhFDvl΂R=2O  BbrenŭQεc]<PaTk9ȶJʿ]i+=%6umɋxH, G"ܵ2f:>إlQ=}Z%t:8r KTv%86 -+(t"e4H;fNE a\eY Nt:NfIɪ7FUv#jKyWhWU9Hm3ӠW2fCzib cʨ&eZZviH6UA߽oBx6]|~@fgҏJW*i˥ v,j=2AJ*2Q$I҂rQP|y0t``ȝQ'oD*5z<4DrRD"U)gsiʦC@Y󁎖4ㄘ|8?Trgk>qi{VXDG^f2]B8c6)~s,TBɗ̍23!BI:3OOb=+C y0jR4LZPDb36'#1bN{7E%0A)dz{_Gap^ Fgb>ɝURGz\GPT/GmTo'~4>Ql4i|0 Rꭍ}=켿^LGa4(W~?OzBIGqry}0=͞BH K_ }wXFaOY򗷳9cI/z=E{~-P2ƪIRF5148Z$2<}>.x]nI¯t-bb3؝y̢H[D 4'HU$E,EXs"3#M8S*}6 /Zn349ҝԼWt[qg;Il~7˙K/X~Cf_GAHlɴɖu xF9.aLf'v_P3h7HkWo;.]lkMW qbhy3=M' jzLr3\Sv3˾_th4^oxЊn/Ѻ3n*d7!>+fWxvY{j;6*m[pC5zSA6++*$$/kVoVHC[# ON.sEKMNn)Ɋ9X*_gƦ$b9O5 /Οzl:z4ݸ;ח}Q1ɉ}G+Gƅm.p(* n=_x Z>jeqz~g+]A6 W 葀* 2)LEKشUלI'*YѦ3]M $7O[$ތ~.TZd[ܔsS,r [rtI9y:Ur=v$^}6سTJ%irulBB5I$j^y49P/%RV5 8H){2hzQAJj彔YT|jB-Mޖ@I:p|9+08;>琾=.Xf ?n$wv |D?d(hCƃvѳ#)QgiSɆmyoψ* E9e dD)L"Qɱ?$=٥͠bx`K?aQ_鹷EQ5e?g-11H)S6]$Su&i[%t-O7xr+i|׵ʹVFL>~jdla6Ԫ0RTWEᄒCơ'VW>.6 5 ņcgې }f"OgMF2jlxf[Gx;D#{!g4PyvuS\\jQm8BtJ+8}pϴ'%|}t+,t No9TU6ڒҁd,4F™2j1GfKUyᓗ$#u*91,UW%5Šcd]Rv$h8Cў|~2Zݰt}U s~nRW^ʆm0(=lIVf'F*Myl@h{Vb_ m.֮WXorIݸm~Mɝo/kWnև|~۫-~CmvOܸD>֊=3-\{ 36nTsKw2\"eQ|x{qUWI,KИ:s;rJ7|e=o)?z/$LDv~ë~ZW?ȟ^U4E_i' lvèxzq' $is+jIRzgnI| $BCI YG;s᷎zӓW-+t#8ШΟ~/q|=dKk3ofnǹ<]VEoP]YKI-o.&?ӤW 7zhz~|1'bgx[AGuWXŸ&_j҆p R%tW_J/]5qRUVCwWMJsDW߄{]m0[|GjkM>bva{+24t=KTnMKU@9ͧ =q^nO mIi\nWb(# YQc>x^m~P cvTZPyOZjmsqܓm&PUO=4i0{LmTj##rs܍{4ˀ!5õy:#|”_m0bGS਌>Gz]"ɢ)EbR>Je&]e^_sFC?h}]ۓ  .b˰[~.8DjY -AnD *kkM$ m$%,( ^ S1<(-y"YۥzHj%%ϹJB6sHUb,I' )bg4!&Z KD(`ƪf`JNYZ&,؇Sºl;h2`[22ʲ8(I !Sn6ZnHBM@;JTU0+oYSft4t,zWs.62e,eWsmὐ -#vOaK  "=Y\Ә+ |RexӚtp!AdHOɄlYy\6.aw9(lVۮjJ̑#sJr-^V STBf "lku>)0I %t'Dρr6#Z)X萚E6Zvvrc {VQb9koDqˌ64JTaYVX+!KˍQc& ˠ 0jJ!BQfEFt)IVF*ejuX hL(=%'HkQ@Q$[8gVl9$8TaI U <&q!<*dKkJ<]4pU>f"" 0e| f7YJ`8 "6ݦ B,", FR/(߁*QvW8hw""x"_3U.ի$B2u.N[ 93_e-C%5Ĉ*۽oeڊon|2*!TM mU2,˨^"Mnv)sW?*f]9k)bbH A&!?yH$@ϑJ$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $z$ $!0ܟ7.q޼qH)C9@0=$@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@ϗ}"@`qA}!@ZÞ< T7$s$l BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B@0r'bH ;@Z+< 6QH=CHp(@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H=jj=İ7k~pvRRM]7i3Kz;6~!0s\*2/RVzeZcSq[ԀK~*Wetv8x143I{0+|0 r(}|hOO-Ȋẁ!CnK!oMd5g퍙. /fHS7EJL?C3m I\sUĕ{Ui}H-gh,D=2W 0S6ľ"|URco\[>zyο?J&4뢖p8}qptGl덕ŤZ1I59Xrl͌y_QC%T%JU!]|:pl>YZP͞$Qh4x17? {?~~8X4ӋE e9:Ii2(3IR&OJjbQdњQp/F3ן&t ]^`ޮ~8x̗i}OnpctE=Ju3_M/J/QnL.gbp굳!ģdh+sԑly4)tAo7kQEwǗRK"׆4nU(8Aa4oW&;x]՝vz8q8]{C8Ï4gML9ˍe/6 -q׀6w6eP|unM\KIZJq&%Owp%:~͙̊yX_F7jN2]nҙ U{+ϣ5|8OnU,-9Xu(R\dLV0˫ "Ur&Jْ%5ƭP>46*fzywe RJ+\{:~>{SVe&ʀ ;vz3?Uח|m\_Ǿ|_SmʸR릭K%6u'TDrq&--Ĕ<7s 2it1Txf 4|N[5r*lWEy.]\K 9f1Gu;ra:mW'vt_F= }1ʵ75|-NW4b\棋NlmmK흅Xi{N7]vYa~9S>udBXvgf Z,'"{pWqvH'u#H2R;*ґ ,+ZjK`vJEZ uUSqa.ʙhYPVR%KI2#s%udI LX5Ih"] `Ũe GF0 AX8$'! ˚QrYܫau=Ĺ0L|h[hY\ɇӛjք涷8fh- \sS3o|S{$»bHa̩R,&)  :LI#:S-uy0ithx1<4iz|kDMyP) L)}+9%̃q*vȽtrGhyby¸0A q-ޏԚ3#U^h`sC$D"D KK)%Ff~`Z?p3tPwrFfw'#^pw&@UOÇq,4w7͎e&niyo0olߓ|Q3R>K׭og%386o^:0Κd_?Osᙛ79ul49G!UIx^ráLO0ͦtRMaьki}i˕_E5F("Qzu]fQ}(cC!O?pe&͏uȚ_޴Iޝd4^}i 4M Gq@BY1UV4rƄ*Eb "?ϋ=qӳF`v*O[6\6MVJE_ԡ᜷eo˶gtcxji@6)M6+Re\ \Iٰ\zɯStc7 :ӭtُG`C(I\*`c)߾V80ٝԋޢ_(uÞ)t}:NI)7݂{tu*P`Tr*o +?dF!8 ˣz=Nn8nȟ4)DX9σd;Ji@윛y)'yKm (H]d'w{3ﶕWvλ)`rӶ&ӆZ_=\"(||Gy>=kaySch4+("-}[F*; ƳSG쒫ԇp˄YsRqfN:7^NOWs\Aw܁w#x蓸M5W;{K>Ҥa](%0`MbTpp]ͷnnԇ5-a*_^3MS_"]MZ&7Q곫Ifj}l[ f dR5dп/]8M65*JK|ZīFtuZ=>^Q墩Cq>|m>6rq3{%ɾ|W fE]VJB)5z݁AƁ[IFkaE6_W0o{lY.͟Қ/K2oo2՗gݚ&qD7Mڼ|RM@݉, M8_AY7: 9vkUS7M5ױtͅk&;o!X%.w=66U%_'+UggNzkʫ-AN29X֑d{aV*$^*iO==h2Z3wDSRƊୂq[;IJˁd)3&̍P=^^O:-g~bɭ@7+ b"~'orQ y}Q/RKjI,Jh(I7/U| *Zn܄mʺ:C- (*f2AJRk)Y!S!өx/Y&cFcs!SU8c1eDQ'c塼 [ .ͣ֩TT23C7Ԧ༷ IVgs.{g҃?Tooܒ$d?5{n%<"Sj &FCC`2B{T{^H/- 1ƴ^zIm!S*ن$,;&&G ьKn&&X*`!1Ṡgzȕ_ y)b103ž`K6,>g}eǒeuGlVbW**XVY6%ZwxjFSJ6d!ZȨ|¸UEQ -HO_;;^OWk`x\@ֺ$G:%,X%>fsPT־n~RtF[Ȍ %4!L"YE%u1u)Մ ue}(;{jEOgq A#(QyQZ*6VB. |6mSBɔdd@,jDBXɀf bQꌠJ31,Y\JlI+@P# Sֈue&ĆzqY^E9A/nbi_(J .:d#2 PXSDQA/B/rjvUP^H'zkobzQ 7D?)wE<ozD3/Lj[W05An6$ 2fs!wp^;\pCHծٞE79` zҡ,Y8]1bD)l2IIA`V Pu\FrA-MDQFqf]`3)׻s]κcJ(6qWJjY#/=H /.$!8lyKF׎M6]h\.; t,κҾᦥ镰.,ʰHdde_2bi 21M -¦j݅'w.4bN&z[.,c&%X-J10jE `'mOai9zn;fgWdO]ۅ tgS/ m2;cպm;/%_ ExHfU@vz3"Kg42TVȍGMrNC^A0C3ȝS7TJe|.$}jgX .ZI*dJ! lRDBP)E. e„F/3dk|n_M:j8 =h|zi2vu]Z4p2=6rOM~n{r94/jߟc?2Ͼ\*f|dkmzIJerg*wi /jzfmNKG{ HryFU۝:Z@_D ori*X(9)pHo}Pր^DYۧx<@b; -YKi=z)m1آb@n ZJ?7ޫZmѫ"^d9o\H&x'![  A/XL`A Q&N=>U}@_ݮC :.4oM h5hA iJEEa+btVQidťlDx{tx6dyz+C+/7P#|;7DzodkVn_H5>qbhlЀ B&(FzQ9A8;t;Ԏ՗Wo^Ge z xh[pR -XRDHڕ\31nڌJkwr-ْ &)d9BB1bgy#ni--e=gM筑Rv@iDŽC/11({Přq(]Oۊ݋ UUwֵɇwfaw+]]O_asz͂}[koza{Q]/YD,mϬܳzQJv|ԝ-z~e@j3Z瞓Eǀos, /ˍp>K^ޏ~j/~q淖cyZ 2;wl| F?h5D<ܣ_ɈgnjU^_+^=J[OUi8LA-\h>wg4>)7 K#Mh,8ˇyA-]A-ףZ5X'A2 o)YD8Z"l)w!@EYCi@!Tt1S(:נl[@Qp&B(wl/w֝S1=Z8esאE6PݿxX=ofyvѦlXGTii BcdkJFf3&톍N6FWk|7;9>ݾWUTthg=o[Wۑ\mz#ÁTwąeqį!Z\ #ޜFE 煒;>>7=먕RAP0t%%EmM/*GW.+k&Q|f'@GL0.*;Qjg١\/R_I~tZD lI 80H|Pbl(]]&aQ~ AI鏤DYyCVEBwJt!h)'Z]>ZY0䛯P>27!Tn= uFD"t1IWB)]!) ^"PL7/ Pz\ z8o-+T"~AHHJXbPcM%*`rl4Xdl+JugV nq #%/UqR \N1g]}h Js 0M$h3ZyC׫HxM  zȔI1؂!Y' H9t{0#wedT!}' k{qnخoNζD,Ex}-itU.u5֓^7.::::<:ebv=%8jec~hQ/q&2I`|s/Y(b0dC$ho2"Hn@IR摾@xW;&h Z?Kg2]4PtvF{P<(vC7CkV$;{}<Y7q1lo~zGzՠOE+$W 55Q zgzgze(KZzg)X,RˤjkFRf(%(5$rN@$ DU 9rcN!F/H"rX$#JJ ^oŨ.Ywv^>̾,ʆmEշ7!Xu8z|kEႫ47US;nn{\,/\z ~q ?Ź;ktI^ss*~zK7;nT* nn歷_s2{᢫6㋸}4 rˆ[6ٺe W~h;\y0r32B-:Ea.]]OZ,\{-|e]#C6$-6NڼŧCV[Q=en lsgkYT.nn̚0qi9"O Mm jD៑1~?sob:2žYd iFdE Pa5p%3)V4dI[ǥ|d>x3yނg<4Y⑂Q_s$=s[Ȋ삭9{ ͹ǟށJ$I5`di|rEW4!PwG$e9J GB ^GH|@%2q 9QJT1m ϲ%ɘI>IG69!1 g.CκC\/  =c~\蕍%0|K6Cׅm=0|K3U%8  Z%׎ȤHha]6Nzh>dmp"='# @&<49d%APkGBD 5%TKFQcs*)B얷'YiӁmx\nx-kU4 `T'#Rby(e٪T C%ЩbJS-yXjmN -iJ@Ms%,R Q$7!T?zGnHIye`"'sf[_&,d"ϘHbP'E[V1y=R)ko;d 1Wo*OPJbSX>#:k!3eV!>`GʘR:tL?~4xw1sBqt7 8|k?'tYUa݄Dfհ<\ _n>=^q>؊J Mr=~Ym'ұ1P2BI:'glI'a68?ݲ:o;+(Ϭ|! Ah0 tlw'Xu7\ YD})u1 }j=Tt? "Yo?x -o5GNN?h5Ѧ缄(Gϴh=75uSm~Ŀݞ/ΞB0WŤ/cخ9HxMzwzt&N(7 wt6޳q$W`s6.} '",߯zH"9(,[3=*M$؈>  &,|8_ֈ&Q Z=j=e.ŠF$K~H<W̒%)>ٰ4c)ArEh}YoP\otû_NN_x2yr:=ɫp.S/ӍAm3?ߴЮꡩbm]κBxq)rK}`9+ڞCћ9.磔z$g,?!Agm~^M2WuQ@5bz; Q C',KF2yhcԑJwғ#e$4w 9ǣg2H&Q j QRǽtsx!o=VUe(eRn$c*ȴD##,bD&{N'~I^6$ߥ9jjk&PkKCG}#N);k6|ɡ瓞儾*VCw&t0]ӄ2B ƒo >@lwŢԹ/<;x O^ETћp>_n>o|%X#~}b3h5mo3(yzD¾( 4b\g!xYֳ.Ge*,yA:r_6»E;jz5gy8U"y'M߁Xu-_?+zkt^mV$odG0@vuV^(}Yjvv]8̀KfϽdG8kH]QRg$ĨXCxT^>r3aL"_B e-a|Af;є\m+?|pJyk[Wz |xypp[`vQG")軔S1zpHsZe9Z>tW:Uc<錎Ӛ e & ¼&BPj@fTQ͍QmQ2)8|td]?*~ b6(Dʈr1I6J]D8iC +LX I91R }72da܀Wp/ڰ T{hIɶ"`5);n>V0|0m'<8RfMzX_M)Y:/;ܨsme1݃pӕ,b Z#@]bc5-jxFvEyu(;n8" 'NR`$',C#8Tdl}XccarD78!\jג)E 07bQgQDWۈbY5(U,uMmB7ZN˹eteVZ(ڜ-rm|UEk#DX62/Mf5BǷHWdAxf1jr 7m*NobK\ImU>m#巍tmZH`ElEO*-hû_# wnF-ׯwV{b۳6"id,rX6c-;׏%"ɪ3\jk`͎}?aZ8ƌ@7iLe-ozOyP8`6 ]ganUZL~ek;[yUͮLkt02$X7B01;غGٿO:_-d[o{ |m—0%b/(IZI-rcgyz]NX*t+_rMl(:.$ +Vr] .Pzo3tV4ΙheQ.oˤe:2Tw3P%g 3Н}!8zi,:;nYU; 99erROMJM3)5MV}cM# ajJBȶbmQX̼ΪΉVެoGnJƘ!Z y%;qT*eΤ5(s3Ғ |犸*.?gRB,f.RGrSDd^L43CQ{0rHRIpȅJmA7r!(8vNxP'Нs{ :h'( k40LKjṕԬ9W]iHၱkT V09;[?73\3Fg^~83 \ @TB yǝ [kD "HN[P!fH BFjRz\TknXe; tN5C!񩛒E$а89|pwvr;-T _K&Z.)ޤgXe(vEJI-Ɣ?.6#E$[D mM 彨 cHGMG2A'LPrOp&L(FU<'_ٴSZE1Oa9$~"U"3ہkw(wa/G (X+Q0 N8"QW1fh*5I;YLG eZ30{-#hFs+%cEgഌJBLǗaor]_8zz}gG&i8}&Euuu z8}چm:Jw\ d>"4&:IM bRI{uyzuNgmV=E2yE]=yr5=m߮܆An5OoJ]ZJ`jzEٌ4M-ܧGO_O^%MdJYXsk1u4mIn0YϺV+aN(Esfͭ$G~Z-{}}@:q߁tdz68W< d͏Vxwe(A:J΅vrgUJ)QmGsxtdXD2 -JHH!̣c3pnWJs]L׼WtJ+'O2G-2}vayι?}عŇ^y GfD)ebm?es$@i.#98)[TŁm''''N'ƞzj 9D)9Ĵ&D P*bLu<@N:u4D{FY; ZDl ha[0 ^ر(8-T|y$@8=ՒPc>W ׫t>`Z0O? Zm*!M!f/>r;`MY'QhH_90Rd\ar aC;NHtk]yF &x. уKCqH)2}>y͸>=Yv0tɲ\\sEOvB[JaQ"@!2ϬVS< hR፳͏it;"EE'҄,* TTdp0<&LnT9oV)$$K(b8ryƔxJhB&:#E MV=?\.֐MD];`(AW (2q[!ETXY~#u]2^HMFY%01Z@ 5a<2DJ˘Wn2d~Ȇv&>Gmi(;dԇ}ib |.ɖ,Kq)0Ice}Jf ?%$spP)l./{}<]^8B"^{>tb@2 >p?JA)L&/'S`燡](8M!Bp/!=wZe9MTe,,>̰Ed&/pLH*,fտ53ͭգ2>'E[u0Mm o-JʜY vI>/kDOc{UjQUaEgN%?{$ ́|~]b+flXo1 S47(_g7:r/'/_z~<c+#AX$M6YC~\S$/I[=3(Y#RTӢlEi5=UO?|N(3'/~xqq'E[%O,>@Ϻۋ_6R]cˮtJ>u7iG-^a_篫_??-%44yjּS8Ta=ZPlc#]&''@$-CwF@)΅#pHT~Y19gd:I_za yuc=_ 7ŋmW.p5N䷄32щGj,`Th:Euqq>}RjwN\[k 4Qz^{S:[uHy˔UbY%#qwnx~%w\]$C-@kσ8DŽ.GksvJpFJ28Z!B@rT `$bMkI] \ɾNf׿^TGdq8[sN"ECˤGrí\ Rp&{0ߎ$O\\vA# w9qB Q>ͯ#9 N?Kԕ fyEoVCHs <|)gpPϒp&-5NO6ZH)xkq4@*k^t|g#S"(yrD}< /:!>aZ6B׃j8_iY!1b1:(y@ eS;*ҁ8/,+ZjKp%"4, S_'k_ɻ"cs?8DYR#,YN.+kc;%sRZHjK}>C:đ;#Ǒz{-*I=1Aq . Ir>ZeҘ[b2pZ p7"nn&}rvN>FCSТe]ww=};l"[ޡ3^ 19*1'ε!F J?̨2a1j$a'u倁 7t  ^X8(A' KQSLITFQ>c"FYj֜I4NΔvvTL{))<]?zV[}VjM } ic c}.qHA$b^:iڄ V_}qo1Enp3N{9E#[/BJx6ϛ ]Ʃ,.2Fcʑ) "梒X0 ƷbF4(ܘiKrq4XEA2GžI@?V@7f+X匿|ZM?Sz|ytzt8>p dqxMf'8M'gdYTiHOt6YLd{~~n,IQ_^k5A)9bT#@>xu^Yo|l&/Iys&wG&a?>z󊱬Pq5k> E'hPx>xӋYN_=參_^Ʈ%ՑˆLq%3TaG3՟mRRlX3VNAG$ A#6d\'pZ^h8Wc ho=#-:~Fs2÷gN'9qJ3uETΐՠ舳zH:g:/ZFF;_eԯWiX^E-{YW]jduٯ.tJ^j Ul݇#yO"\'O2%O>U g;#< lec *m3LO?惘oTI#2pIKb2ݥ_/ѯ6 _45x xO'-aa͏ohXݠ!m$Ig,TPlrج%>}fi7Ii/\F%}$3wd3vΨ+ ,&4j9_^ &] ԣP7i3#8 %oz:6jVfFimr}uY)M BxUȓ W̩W0#WOujM['2eӚ5+xWew&C(s?`x=;{pT({P[\=qD[5m^gwOT;ϛfw, Ĉ=ݵ^QT ;υg!DSF\i7&\ͣ13C*!jc NP,EyPBՉ($1 ;$Ok|R? g6"SjWYҦL1<D3.1*'Ӛ_PPiʭ 9/Qj1=yQx1Д)\nwXQ'NVvkk7ؚzTϯ?B%QyFE`UiXED ]#Xt` A e7*)1p0MQFꐬ$@Ǡ@J#c1q#c9_ӌmBsg\@,γgܳ^lvWVxooa4 狯U Rgke%nRT),&W+٠M UD}#v1q#J/]L;Em[]o&9ɣq1rSB1*XN bfNs4xg$!gHFўGXrj G(\cx e~eq_~l]O1D6耈 "2kI 1zP@0XuTJ@σ3al%zOBD;(v8JQKg\pވG=FLh Pn2L=}Z,%"-\360v#-H^9^3B@Qv]DΥc[HZAv ˚Nګݸ@h$Sd &e>Mf߳A_. Ԗ?d7p"4n뉇3^v/% ^22@ KTH-T`IBZ^Z,gy z˪=yrU a{/v/ 09S˳s5٧zC-ykg}V4ʤXk!F^蹌DS,w~2o%r:?y[YK3NowP0 C4*:H1ޥ QOid&f u ZS2F Kx 9b I=wA[ےP:^/&Ά՗hGluoe̒ } @hwׇ]/%w5YԑCN:h ~N c" )%9?jh:J.K2xLxBh )hcrCR$J`B0/Ѡ]2Ҽa+c߸0&Aqi_6Q>5Nj;ST9e蔆x0] ыa!hN.肏8%H.DP$hǕNTyۨq~;Eܿw0\ trcfR-wlV;6o#57b;>##(*hyK C86_[w\ܕ‹z1J*ؠ%ՙDj)jo -  ,Dn̜п^,ox#ջMu>n&N_.qhwxr9}m1dWBC*D`,PE-~ڠ 8D"iReZRMJۻ˝.M۰ˬozoMߌOS^Yw0~y 0mӠz&z1vz֐j*qÐӝydD&ߤv9 56'Uzê㗫FY5jfWGAW孲ka[#">WO_#[DM4>6nl[Փ@ R n7:he%;q1Ql{jS4C¶5~УmM&i S@zqwg'ʜyApKl?%y>aƧߞz'nc쓵&$Ќ\c#3q !zH@vhToޣܼ۷p_=f䙳/9]E_|~ֽy"7xyI|Yk&ى1omӎ6ΧԴfhD6 ^ 3perXZ>by|2mF;xzYqC$}ך8 -IdR;dJ~tԃmgSӱcģ|=FdC7e5Qԑt@YJ>Z&}ࣆɚXQ2ţl&8\{f߼:pathijRqhY65:y|~lLJZZZ4K{觔Z0&~nHm#d4a: 8p`CaG'v!$t' C`l#6&dґУGk?o~xW 6#O>zX=-,~x3xC[1 O-h68ci{CJXg2Ӱ.I_}y`࿶Gu;w/[ju+k3MPsrR* D:W Dc3c#mzyZkKW7-?ǹY/؛CݫYZ?O>u4t]c)#M27Äx|NR3~/,`nvnf7Jј%UWuvpfM~)`0'Xe] 8hL 8()PuoHkwו N] J(K[[uu]A@QmP+%Т+ J(K:0YM;JpjϣJ(Z\r`5] ӣ+%5A ŏ] eꊒOIW@(sAZJ(} .QW^RRcʺs =7/,(w¸uw낚us_ݵWCwashoɭj8r_\1gs]_SG}[i"uMԼ~;`+'^`[)_}^l?6D~5no.ʵ`rm ?0[ۧ^0o_x7@Z7&?>.[CwyxlX4,ݻ?zy5߯~r|TéF-}D=_߾itZیMMj:R3A;eڕ\kZ)λ& b׸6&y4\8f5SpSiduơ5}ҘtP%]8fhhoۦC|:p8AЂP IBռP8Ì^8FW j晅 (ҕu:FW;{,Z3E}u${ Ȟcq A@n S FE`rjt%JhcZSQ:Cu]1.}W;Lj"hȣ+E3ȴDt]1eE?yO17p,] S] u%U]-QW)3Ժ-H>)΂5)jr'>.|Ri%SSL|L-YrZíeIsxi0be5Sj4H\9mJ6 {3 iↁOGXC FЂb#BIubOVŀF<3ZfY ӂ/^WLm:R`"]1z3i<>HW)S~tŔXu芿ةhLts*̺%w|m)petefw]_[uO[~R+Ns!h!RbJWآ㪫 bpw] .-ZKוPO4˜"] pLjtŬ}+GWB Tu@]d]Eb]1njƮ6t #j94J)ѕТ+Mt]13Xu@]@Rʸs~@ҕS FENOT)^׫ccJgjzi+3njt%R("u tL1nZtŴ0әΈ~;;V=3;<8<4o<#* llվUo1JS+Ƶswhѕ+,lkwitJp-]BgfdoGWcÙtR_U0pgfт8< ]AվUoD4t%Jpע+MƖ+,mDItd""] G5܀Zt%x]1eq )"]1eDZtŴ)+,mqIt(դ+&FW;Qyє+Z%'5ܹIѦPc .QW#M"Aӱt%z:L˱kbh"u+w*4ƭm8wt&~G?ubjadXqѪZW҇jZ:磦f EWBaTuD]O;t'=7ձBK0e;]Ug&Z[ .YB4@<=ܨf1Ɨ^~bu !GU2^ f_PJ(vtvzwfL0,dU.yuGgZtG [2t媮zk|T+FkJpJh_zGJ2UW XG1(GW]1m+tXu@] 'EB 5ig[tGjr`Oxb`P4v%sG[~gP(]ZHxEࠧ3(sgɣMGW$CWj@P8;5[ffiRbufp 1!ŦVgM y hZpNE,i:ŰDM`&]1p=]1uŔDUWKՏ>5P f`vnrW_\^JDbޤӾ<\|u_x9F߷\WuVGO^-;k+z|/k//9~nS!Xy+_opn-7ͅchK8&oo^M_w9rco=7~brKonu7VzwFܫҬ߸)~Su*UnͲψU4}"D3a _<ퟙ?oIG%>~B\n;򊣧_߷̧nauL{.-8žEh}ġ%;@g}GӍCo{Omb |wǻ$9Ag}'_=bpE]6W#k=LIvCrc..0u#B}0"yت?]HеqD3.Xm#v0l m uS0o}>lK\H{a8m7# uIy0 .k`A^mB-6A6=9"[҈= f_OfF8'+aMҩ` .DK &XN7-b0&ڠ-h:hѵѦd顁\o\CRVa8 pJ@uM+5RœDhY{n#b>;Lf 0kD34p'-Ռ iTUiԒϯ=='z=wkѽ4Yti:FsRE4P&]PX4Xj@cV91F<]U~b1*xi }"iɗbk*6YtT:DyKDn $J!w,ͳzM59ՑJj MQ Ar#钍5:3:%&z7'ߓֈscO1Ǒ֑ ~F_o3dT&΄|`575K!RXuݷn/D@K,X']4/V[ZQW!5*dlzaCKU+ܜ-wd0Wz,I KV9uV":J(k! e 1P #qҰI&}EՊX{Buf{ ^M/ĥj*٠db̠JUYa:iB0`߻9;3TWW~9/x|_ ިa}Qu`mFL`-$ >:gAuPyP\J_6#JhIWo#f BX9Y4<‘ϫi{l}o8se, BttsYhl) l\c FZ} \;Z,Fvݚ{9&mF(γF9@[Fm1PNW= @zQF6k80)QC^"$u-ҐQUr#as":E'%\aۊLv0X$SSAv =>U߲Yal+4Ov XW Ժ) ur7prR1TDyð*a‘r;pYTQc$T:9Q\J#xL 7A. lUϤ%hci0Ih j hN\76xk+fnQ4Chփ*H]6 |tg҃d &S@rZx tmKzg=sTm /fs jX*>Y;Tڠ@J+e`j?&zFv/nY7۽aa S\H8F(햳w(N߼zu}mPUT(u;Nk/F3nwW>oj7ח`eחWgv޾wƷw\^\vxn~߫6}uc8a|wSn}'늟;=y.r^`1ۼ{q{~zk|Jl6W/_"T^(Oj|hX_oꊯY]^~y6IOO:N_zE%mw}sk^۶oN[El)'1dGN>%GW;&U"mȸ9$ 9u9LJ WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\=p̀:MGg ~x^ WpeqbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J W2\Y>35\pLcpeMTRx7p%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J Wbg5\9#2\AJx<+(JrGcrh W@y31\=#ÕÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbZpG{ tիs˭޺a[hw] '^T6N?TUAOdM$-ЧN#+^#:nBW6NW@YꟆG{xA>}qzMYzZ]2t d]ЕԡguGDW 8ƣ+J ]1Z:]1J!]h5GDW 8~ӂ+FkS+F3+ +kñ&m:]1ʃCWT }Dt.tb>ub!]m8 ?=Pؔ==`[vKܮgWmof2tܩʞ?'~fц\apq6vOmza,*qOow`zs6Tm*~{kޣx ? nO2o'sLP|nw㾹^!_o.2(|Z_=\-xWiN.3mK>1:/ޘ+L'Ga4m./V?+f++ܞ]Uw|?K͐{pZ2O7Zp2 ̈́JM)8j&G 2(yq3ll iZ[&O)1yj£,~Zf#0޿4gJ.M?GlNBvq _7:ioS4&*Mŕbz|L>3{fiᦣYZo0Jei.-E )pU"re,9ܱE;+2~?ߚ$ڍñ6a64QOVs({5s|fQ_V{ZS2\/瞭nqcwOX|ߝyu-' }ŗr;\ 'bhvO_Ĕ7ٴ>}󡝦h#tJu(.)=Shx (ꖽw#Gn.΍s/ e?OX=|f.ob;ZA&ܾއ'?{u{9 /pMv]C. /7}fd$g-m"M+?F}{y0>sy~?/~\ѶƬ:-e -wǫ(d7hc->q\Ybp+3Dfә)E\/`ŘN#a^kٚԳ*1 vG%̔'s][o[9+¾.x 0K70 &=(l!d'-mI#KJN8Ey_UEf8NKw.cHzӚ*Y$z>$Pt d%&yjedl>N}۲6nO_rk{A0/꼻ߙ=񊑻|ӧu[۽EVr$ӎ Z;KSRHEDB*xfBgw_\u^>b.Tד܀7P.OʡR *,VT.fo zÁӣiOfn/r\dW_=M+:Zq01' 0v;!6I=l0h2HQ@E 4FB:!4cW§PڜKɇ~r@zMwc{]/)҄4(FZG&ף`Bol aVi$|]EOn-:::R 5 $(ih%%ɭॎx9,!ŢQswrӹכ׌u>ja[+ t)z^h2l JBpSW$:wQD23e¸52ǭmXF>MY%ֆǥfg/7i%d^Oޠ&;jWϴ'ՓEmZZwFgSWX#2C1|HR M==|6\,j;۾'}0ȱ >ePm!<֞=-Xzţ6e' lV;| \5fdT4G%UU@#HCbD)8Ⱦ' E2X)k:N>ϥB:(UL$$&IL̨بФ)i2-w)E1+,Fg2'D1{tx]$7l=WB_2:tB[yBnV5omsp꾹iIingζ'w=N> g{ |JYs5vw]zdD `n_j]Qzu5m{ZNmv;l{]{5eE-}ӝ.J0 v7@Zy' iMWxv{z •YX|Flv~plGo?}iNX'l[Lv(G-9XJw8)"2GFᦉ{jP+a2_yb~z^^͙~֑)"SLdF,G\fR R*l:zg! jR0 v$߬Q*8 YJi2d?[_aa4y(}N2}ES9jL9McZ7si %Q6o-9qNj Juͯu*`uc7 DI *82l=H#a1W) XZsTDm@4 ]F"/cZCpt^H|T Fځ`)jGT/t/C gx.74zer &pzu3Y1-^=^LqyNe |BsɆW~J͆dó\*-;9}jValqxdfRkDj1T|=g.ǻ#촒6(adDŽYFd *mdlrܳl!\..SR6/|s.+׿/5ewc|քD01ZXk YzC`O 86MfMr|tx{CHaIM|_dd{kCI)G.1ܩJZ=Vn}ɡ{s:h/H FEA+c`lfr4BHXV``֊ mJR/im271Zz;txU]HPeM #fŔUTY@+d|dOW'^]n3,u $%3\)Κ4$IK(`~:${8 R-OҮwȌS4>۬ 6^p)ƜJd s',Q2SѤ,%\K$6%czv^ PBI2P ;ۋS0c_\[p&94JI:gSiTa<Đ]=;NcWIs #8֭|9We#R{9kdP*¬׌z3AŧߡL;78Ιq X %T)ea'HEiG6DhᝡoAv'r6:zr3ۯ6y"8Uoo Ů7mkɾq\M쏧Kd`rGF4 Ji*d"WUD4Mē˓KN(:x,w $Qddq)" %y0HcύbijMC#hD%(9J6޷F^Bu5c JlL>Zy阵QB1g,\SO_M&'ii -GYQGe)dTHm̕vVɒXlYpֲ:9kNZӰ=dNQj[Ks9<'c:R@ ϲjKDdrH4[^n+q9ugDv{i7ݶ¶!TQI1`4ymq罥ΔH]rg}<2^n`ϗCk{s>eJMpa)r0:.F.lo d 酳)(s1,FjD[0 V@!l8|cbr,y$<,͸^kbB 18/tr*ߧ9]JmLNnNtJawVe&~!|VTZЋho=ޕ/*"sQ[%d:1F7gcaDfE`1c?(GXTJ0G7`m"%@=Bu(#Al$Diʣ,#I%E5O&i*@C]aɌLuxqo1)ٗiüH{^y^H~` -؜$gg!Ԃנ0#)IΑLx^lܔtˇa>d(쁤*ѫ#v5x0qAp}3E?TxcCRE϶(6O Xer*`х&^LD.#?28vp)/푪MٞM3NJmwY6( d (@ibF&f w!:7LjMX\ 0:-7wqV~磶ےشG+ٴE5ԲQB gg" )އQCSSڬLzv$!X<8Bh9jcL6M4sL&$,hMu3\%sZX/&aIhöENi@ pg4:U޹ev+RXXTŷpW,/vGc'3C[-!&H1O$g)\LUI86iz- 0Flz1s|6% {@7&\k}~9_P&#s䘳P۸cVu[5wMRBCs.$HTYba ʓJʇqBJ eݺ eӢ eӊ eJぅlc,1RP*j驏`d#ʆ48<3͖{]׽۔l7vkG}st_QoUN9s#q*! ,%:IH% 6֮Һ7E}6ti*[z۵IU=vU%|m%᫦H f 0"u{uG'U=-~(YsspLaٿ>ihՁ4tR_\%"=F| %ܣoY/(aePmO`/PG{fɕP7mVN8_C^?VmZAn%7Gk8הNV >\u)VXM Y^'|p_~Temp&XϳHjz1 >/2'_zJCt"+kdW jNWrcx [(i!C•+thU QjgI!BFte]+@ˉl;]! +ӲCt9U+tW ъ=]!]I7+XwA+DW ѶՎ(u]#])0U]!`K;CW-o1(% k]Ff$>f@Q-EX49^nnAu`+9|H/ ^|NοcwdAYK!RJK5Uo91]iLMwJ˾i̓hѲֿb@ciñD kBtbt(y;Ff0l+MDWNWSOT<NLWa'R?+P*ٲW z=];T`jk RӮմt(͆H!BFuU]+@k)k;]!J&z:Bs% ":CWڮt({c( uw ]!NWɞ4Z6; +th]t(J)*]o(]!\κBWVȶc+C%EPD0bhr!~%nfx\x Rդ,|L'NcU8rb9 w"E>;*!Dr>nߗ Z[@rrPe? `~4Ou[4񝮞 Gz'6jIEfXX-E4p2Y&|_[7"^x8꫺[qa"!mq߫'I襷c+sF:Z$*r!ނ4KlGϼN)+7s{as>GSDVŸ^dK7j-+m&YUXsU] " z`1vո!:&L͞"Vll>f|D],c!s<G'lZqMK\?յ;P/ܵhmw-JAZ00ޥ(>Np'(g{Bft0݉CtsewCWSoN3np3;凎FW-Ӯtezwt03tp5 ]ZFH QR3D!Bw ]ZEt QJ7\.iWXU tnv(J] 9c++Dg|WV vJow ]!\nBWVT#+e4)]C!J jSvoh 2ow-""i M#ZfNӈRҞbTeP]rN-0@QCYNh թpy: R"?jTg$$5*XӖk녥!"qpY'NOPE#DF7Y3"K3>hҭF_Trd< W{[Ǭ9QըҲRaeUW/K;ص/'s1k[7.&3%#\''YFe/7}]lMD;ϯUXTKV(#7 )v:jRT L1)u&x7 oz|05u&# &763s *Q[%Q,+st@xjxԊ&jqʤA,p3 x <(^r-`_P|P* X# @j heg%"&#!\trE.HV*@Ef"%c.cfޅSPI p?Tz01M%dI 8PQHf&0;TƂN2Բ ”M~;{:D2rWSм ݵzn2Bmdi{~CcŘlnD),ݸ9&!\fϖH7gIUv"dIJ}{eʵU `E|ZeȞÿjOd?b-%Uh>kC! ³WINg_ž-]|p#ar>۶w\i)xOgnLVNғyA! )?S@!Ra1D+R S ߏӭ;phb߹[La4 _{mA6b9HE$pE-! pSI Ih"E'{ hZCi GcBMQ`୍c!TM={z#vdu]!|Z0v(Huz/7R5_F %K7;Vf0GHk;umf8\dVfNPB;cMG*fyg@!v£lqѭ=:Ĭ%'Q>yi(!(w:jBMPEY/^ v8>kjRn6@=o> %V!"tJTģP%F;!ZXQZtS?#dؔ\MrT xf}}^a--Xky$qe`ׄ#j Au~kk꛹G|)ק eF79n dcF~3 __h_% 3bkw9f$[b[_ϓAf~_l(vc[Ga[ b+=ԓB -R_7? \{p;8IE)Qp11D yؑ}=nd<>Sd$IQk41+ɸ4@8@JckW'pJ&fZ>罥Behu(KYJ{zvFΞ۫n>$ϮQ+#W!Xv8lR{T<ꪚyu73-U/VLu~ _ŻIZ5 k..UZ?!eKMp-.v؎G~hbK nݿ]{U+_أ浒a6쨯}@ZyG i櫷Tĥ9fZwDl,Y~[tȐն'2OE [ r_܎3J62γTR8 / ?=b|/^Ȟzԑ)"SthFm uBD4r0\bW[#s8ΈMZJe2~R.%f :$IM$X20G٘1TѳQ9׵!Y<٥/*hzpOC0J5g Fپ3AmZ¾OQ}yuZMp\{;"B !i*l0>EHڟ?an= 4;ƽC4YE  yD:!6#U$`z69f#mIg<#jD=2:pBFQN|C:x FAʍ^>$h gz.f7ylS|G]:C `usīτWKelJ-)BPYNDZQpA)g+:dR ԇE""Ȭ"<!^fDJ$BJb 81AN ri;|:혖.:wJ"$٠G,82 c IDA 4,ȑ,buv7Cdl_ϫaW10 MU6L Ec*+2: YPcO.=&.{d*2pIU`p"hJħp".M$EHّ.!,H2I04FiF*Q 11^XIAJdа +H]v!ь&Qd[{(K @SJ|CUN/NAc ;eWv2%'p ꤚ'Q^@ d.$7^ဌEʆ7Շm#nP—CuSW+(~l^<* hX|\Jamۺ0[ \+ J5FM^_k7O͛.Ńl 1lqigM̹:*xi،-]s ?^볞\o!]-5޷$λaX:CAPNj#G1r|tjst>52ն]v*{򡕑ܰDdV_pQJ >'t0i\7Vʚ;*Uh*AMȎzww>>G^'9|LO#>Moڦ4bh>-t_mЎrm/*wqzr>͉]ٲZ5 ~F1?_^UTU*Z]]:rac촑6ĝ$rY`;ɀe348LQp.<ڏz˒U3fhFVzN' @mC$(sGjȳ34D'4)O3NJa츯˾Nٷ3G]} ĆiXߣNAN=H2'&:cN0p&@s;rn\ɡs vgto3џ~a_l7qZs\N0-S|'.1[&!JCΤ}vM9={[;.@{<0"߄qaɶJot|4+W""H A-PBY#x>ˡ;4ǜǜ8zqN$Eg|ǩɂu)`-#FA@%!{vf<39}%/./1&rNg7]wZo4vB,|ؔYO6|>tٯͲ<%dr v JG⃰Ȯ\jٖ"$8GL )e)oM.˵קk-&}O831}AR(UCt!&ю|$;RَP+?CH/rE#%Y]hgVY/4Zk"$# ꨕ&G1zQ:$m[|;Mn zyT[;7֢/ePYtvxPD eT#s%ud 'vME#d^A~T Gt\ ͞7 CFT AX h $7ʰ!AJ^ը|TQhH$Ȉ0KmԚ3əy=b,9A]#^G5~<"&Ḭ@MQkN)/|Zx6OS#]V$0[qsAQ |ITK-)'M F>*~d6ƣT=G{<&*hBZh]ks7+]NZpTkkdkRJ&i].0OV7z@A'B'?sV$)rJ3A$O'd^8Cè/6{)ۛA ֛Ś1Qp&P2J{vn3ERAJxMw ]NTS.29xEZfOѫ:X *N?wW)N7Uv*R~X62RhWG' *_瓋VTxqZ}2:W7dmOG݌?yIFMRol A=]|:\qntܔdz _oր55p.5{;P/j;+5ǍDC+VN:;ZW;t+yw/Va],\RȕUublƶ+6uD&~nWZ!*X5aRϧ6niy=rv[7 )獒 Al'[m~bkq:kQ4k~] ucM VڰPGW61Z9 R#})w+qh2 հVn\Σ'G=ezRH m m@]̬on" 9_6߮pهL.GIG #hw@d۾}~33&d{O}9>y!uݛKƳ e޼g5]te@ݍ{;/|{0'l7x8k5Edzdx‘^Sve Jk 6t_n1s󳓇'm g9^ I p"F,P@Z(3Z1XV, cId"(=YF+^H/ThӬ+ ~ +O:X8nslNokϔjop(:,FI#}D:)$=Dž62X@(rZ_*q$sJm DlEE&ƾ0b##vA?q _ #2c,h9'UG%r!9n>P¨!)DEBGm(Yd(láW3Gn\ RTuM`ntT L8Mre BC>2ʻ\9) 8Qpz"s̈́RԢ 5iF dL lCBi-lڣĎWv:Բ"^'u˛3۸+z;BE5 *)l1V|(-r&bFY#=C Jk@&N&V^vI-i#`Q|0CM&XixV*ȝ`%T:kdacw)Ms|D !ڧ@r!jGV9*QM3+D`1Y.xh\Lne_zΐRd%UҞ۰.wgh#?<3,=œ YEs{6O|gKWlvfKι|z1*qJ%wΪ 5j̊hb8|BGׯW=w蜾m_vIqowf"G )T 1P.3;c$:@ܣgpT cvkjߔuߔ\6ݚ\Qhj4Ʉ⸚}y^bc Wͧ#7<ޞ5}6?'k̋>[qJuywϣ›K3m١VQL%T:TdR3hbiRuJ|rthWd(ķ,Ȕ9j70-o˄" w4^= bj|rW3 &+Y_a%IL"5[#kcCКM3` ;ގ|ph>QfԮN%\~EK?$]LWOBʂkc5$DdprxnAa7,7h&ģv?<ʹȟ#RCBv^u޴^#/?wY]fMFuN.;XxEް&|wEC-Vmt/>3g-9 ?TSkT%{{t}r\aoН^BN@ޘVoGܹ ewro?]b 5BZR@们9)jw=)5+/w k4 %2y^ydOKh$!KRTR#'NsOB]^Kٰ^u 3oǿwQ{C)QFZ%B!%ANa8P &o f_W&Bn-_/˫&w nX7V oY.0 R04 CcszeJD )$`k.-a'A S>#Ϡ\X 6hj1p.@O HD,϶$AΥҒBP ƐsC9$I3⩣FksN"Jۈl B@YX,}Cر[]c010{Ȣu+J|"qGZ"mSrKety"8h 5e+h` Tj%ҽ0@uȇ3yE J8ꂱ&i. FB"2%0ewg2hϙ"jMdHf9VC:鉶6S-g#Sg}}1|)OprE[ս믷VZRT G)*&*N(zO@qLKuc@ aN]rΪ{yu]?yxazв;jl-L1R:W8T\PɡDf♢UͣP*:DhmSBAgQp9yfzZ*MW-ʕteAWf]> B ءUF+H*0++!UUk2Jzt)0DW<g0 fFr'Q]BT0&2`I2\u0tj;]efp_#]I8=$*c]eB ]e2J=,Jyue+ f@Qו&h|vZ] h6Ϻ V3^q*_'Twt2oo 5G83ȴbSc%$˧ӻۀW3;ne5.q8F4߿089&1],8Eף/)5Pk?^+duŽUkq<> Ӻ7 +VZY@;iX<-jxw_wfAi]hWG/wo?((IKkkkA*c\<9Cs>9)V)%fi)O5N8S'&r{TOje䠡KpMR42' Fhmdw7 5TZZ@YZp',&K +\ZЌiJ2`}8Q|u(tB3ʕŇ^]&U,`*UCW}+D@W/l@ &z$\0JWD#Q~K=]tKSL+  ]eCWd}Rˁ^!]1 XU+UF{o]etJjp}p|4NN0B?~r>7g-4Dۥo~~~tZa:cY}dhVvX4m^ -PKx%Vlkosb̫8UwvZME?T'pvbxB Yq "ܲ;5<ů^t5PW ڔͣ 5w ;~q==XVN<ϝ>ShEEܨFs`OOode=XV}:6>b/XUlqw-mN {sҴ ТAg,D\Ivw%ʖlI#2f1I~;pw߆ٿ.|hm[UV5gmהGM0̘l [-Սw-%Ǵ՞ƨRZ5tmM$,tP"]j]%\%BWUByGWOR 'LZNW ]=E^a{=r.A[8'ӣ)uϏ= /?!:m?>@f$San,s]:_]Gʑ=ϏVDo FA)%D QGt|Xa.]8FBdS}iq* za.' 'K!қRVn$bm %JƴIwD{$;D:6 ֦%TĹ\BBW -MRvHW:eAmuU[c]%B6J-Pv o98LJ?"i7znpŹwCٴx];+Y C9#\BW M+@ɚH6`ni ]%ZNWR!+*"Et3ꢫWFtP|tgO #YW 3V7޺iK;*tŅ"MvLI{B .mM*协]%s"] h3%LhERv+(EM]O2zH0vIRL1߸vk**wlWZn5L*ܸCwkB>Ƽ՘'1g9nAftg5݆62(Qn;`2֦HhykB ּIhI'Tw'ZP+&Ux[*ϜPJ+-GEtIpek*ոt(^|;tEzG;c Fh0= wa&$p3 =?:2cx ~flv*Qib]R eFHr Iw@B}(!`.Ӥl?z߽OW#qO6oF(Z# bDFp$,0*8㈠\he& ɠe0{ ? ]qn l AV䡿$OgG X[a33p} e{(Q+DXznp+s'PvC`qY5%-b7)5lr8|VZUd7M&aq/)Λϥx5hAf83G@$@wWC,Pdcr8_}*P6dfQTIU~J]ۈР,:H#]H!9͢?.uZd 𹙆̇㱻ȮXeM hbQG"`"@0+SDWTZ{`8I(ҸW^#SuҴ`;4>;]֜r4eya{yq&_~l_UӅ7.ʹ1#.tx&ϟ\ѯ5%ePVys9Og,ŚP}LDEqSV .μlo侣2"%F11^XkKв0] -[ v*wL)V#OmY%I]NW9( okp13vBIƆ#~V%7ih 5d(§Lo0C6?x7{rEޛּg-ܧyZnٺ"#$TaLGĹAUՔ2T[3F(C Qr.;k5(QmGs+:2,"1rbrX+_ʙ5G~L"9? }z|\9H\Rs10fgSm ts̍`gjښLq>6dgOY6PIH8vJd&^TJK:{qL7;zTӢ{#ũR)0,Pk!Hi4!4!ZtR) pR)!T7‭iU"bK(F[ :aFhNPƚMo yI_P=7‚ۛc) 0O~lj=|h0btUP%N]jhI)|YךDDcg@1m#šeym(-hǂޠ[yF &-5\$PPd\b9R8V+ (WKM)Qc(0(A@)"Gg5[rAĂM e0@D6(og~0 ]_En/>ݺr_†aY)\¤(ü;wsNh C3, Bxu>D 4P(< 长ub%T*:\r"72tH"9^i~KЈ`@vXqYo9b`wq}J%e^?Ii}v\̄9׷i%CHFb<HqiX32`_G)(t? Pg{O%P0 $+:mqk+K k2.믡Ct4IQ͊âw~y*!ǟuD2%or&:9YF2x2Zz<NnʲuMK 0[4m 09]RRҦӽ^_0ŝ/.gg /09se0dzo^m%@6233Zt036TdX[k:T \Mzm5Uay hs`4G벣:]+ꬭk-֍mUXE9ϏIOr {)|7.Bx %E­  F\8/oO_W~WoO1Q//`,8 ``? ~[U&j6U:[=-ꥸ&oz7j[sC//Fԍu >^,gOHA| j~HE~J!ܹ)D !i9nr+>Ɲ>Ҫ5}1{~ ;AN3JE^QgAAaIԂZBT+Ŭv,מ?th>ʲ4J  P-F2cLI42"F\vh4^QjN7hbݪY-9I%XmM;[Э8I|<'-9"^̢w9ʧ:^z7v-<ݕZ+~o)~IZe'b|~6يi(_ɽbsYZo/h&Ï/ԵXYmr5fp,c0"GM,lMX5uik3!뙤p z gAa 0h;@.`D](D\1J/?#!fK1Ey|G2+eV6n,^rE+}VA[s5S$M:9-e+R:|+j^ 9!\.ޣdCM/ 1h@#|iqoXz1}-^.Fp/Mr1)"R! .= . > =:FxFH|3CG _x/TkfH1S 8U`$8D"Q] 1 k%,|>KX{))᠜nb΂wDðqu' Q-ms9gvҫ)XEvnBqK˹l^]mh*K$'ߺE¬ID!F`#cLZxP"p_"\rQ"EAvS'3QF[cx4i`$ҧ vO(< ȇH9scZ' kɆPP8$#J#ZZJ G瑥=*`A\QwZ_72GFc>WUbnka8$c>P;Eu{v{?Ad/AI.ndӔys8VDo FA)%dȠ Ap4) n$C͎w}6N+{Jv{J8:l`\vb]koKr+?;VFAd7EKLZ S=3HIe) 8ÙsNw:e5C߯˥Χz;$/'9aZ6.5-g˯۔Us+i6 ˏKg|7ntٞT-ӷ.,}Ͽtzz'qoC0uҕ{a%\r5mZkTw!g۾FPt@[ʼn1|L zUqSY-(V{cnG?S.o_Ҫ}.m΀?)tƠo^\&ۄ|(~7kV&]W9-Ҝ䎲˿ݹ3>n5u˭q=5,5|{ *F<`E*hiuzfp6nob"pZ$3-*}i Y^/Wsߣߝ\.N0K=}6r:6[9{u\\|l?&E/f=/?p~C{1:'>nW,&l/Qњ7l/}"/FXm]zgOH"=1*Q+ٔ0%K)zťw! Yf.s=:N{cTyd.{Hkx7 -̔C1JH(QKJ3hӖ^+6 Z$:Pb։R̈́م"(dB)O %e*j7 iQSpq݄PXj~Ylnhwa+Uo{ɃlZ\w=wM^FE3ZP/ӰI6zLhʔ&  6HHAԷu"ݨ0`s{iBS\Ih"bBF"Za&'U3V3gUj=X_}!4UA{9n"'eXM_~wBbzd<|tA0iMr2̌AxA#\"8K#ZVdV Hx/(nSFE+\p,KʙRf>=v5s{l?-9]w]eF="؇zXa@2GԩFD{qENJ׮֙]t?2!CE,X*Ȉ$dKT'XV3?lI:ʢ8ơxjHTG9ZD@N0Ȫq 7Dۈۆș!pON!8 v0f"nA2Z%r3>*d B &-h"j/fac!uV%E^//~And ʪ@Y0rF<( "*OÜkcq.@RTΑW_6+z* 1AX.6я}oeCH+ԯw|Ѫ&ti}&]]wʇwEO*ӫ(Il 㧿(Ez=۟iւ;OVnZѸh-3LeֆO|u;Z8c&}~ /"E!J}%ұFhns{7k:AE y&K-ti: >:0fu-;`_<4XhMv.J2Lg''r1B~y1?'4=8ak)o]l߁O995myMJA42y0ct7'/|-k>O?.Fں$1+#W5Ȓ;u66z8O8kYQ 39m-ǔ,JEJzpE&J&&4udc6_f#ë7t@Խ_9*ok,\&zKCm>|n]NJ~\UlIhVE%cH@J ,ySps^w_{!8I))Hƌ9R# h>Cj|i[i5X>X!"pO~Si['AFnD;crdse$Y+^ .)Tt6¿*46DVRr~XKF Gl;6<;mq`ghX^fA0δ0nx3zaѤf~+(jVi Q4J8UB=*ܳhUT~[]_J؆,?~prBU29Q*IT|\bRdO; (cPEtsK= {K2VK)lz5cQOW:t&yqqcMVt^E'_Kn黶٧l(ߧ}}|$Rįʱ P[k= iERJGHBp/h\U-:QcSzˬRF3Ø[9qQvͯ^/\#QApA->.^ڞlGBHKhNvo/W[)k%o7 4*l΄OQ I`kQapIeO~r4`E'A`xƬ2qP/xgeTu'.ʞBcs/lXksoM3nNyGK!8Ιu,ILA,,+p#\VH׃fЛ{o^H-N5g˳AJY)̾Msw+Fd"7U܉hỶa)9z՘Krze^hA=*4<18.eVEU$ 9\WZRk.t)PsYqRM`{ dὥds#nv<A%*˦$>7"n %1ԭ;y| '#M1=*/n}1} Sp *2O1 8O=p?LDS*GA)>p!0\+NI.YƪS(e4FIJXJY!1!:\8#P L` ? \.߮>fŗވ>lW8wwڹZ}Z|NtZMϿ<&%TEu062*&Qe4I0V:Ǣ(ap\g Np!'Br: >q! BiPjkqW3gqWy_ث_?6{T R~^12bh,wkrozr rTRV#@K7y(b ϲ0JH-O-5 T\fPfX_d# 苘a`LsA0PF(x$#|RE/iY\&P[սy->FG/  2e 1Br%4:ͣsZV%mWRȹGuo`0oudᐱǬԌk3]ik&Gf2T2auuW޵dBL6mb`d',3Y8uS$MYTwS$6IQE[vTשSu.|XȝzUvƷ1'25~E'^N2cM˟(Ikr{NQ([IA#ťT0f"Ѣlz&s(%E 1'(ws )K$EL!9tJ; gOV^FWXj ]D*}8n34_7G]FpKl- XeSM:pt\V@W7n5//[~z=ᴏ0ln:On#sz{66z_iwmo!Vp&4x鵽;_U.wWxw|w޽y|zGl?ў4wVxgin{Z~ywI-ݹ]rƳVWoաCOX2;kB":uVKClsk]sD6\9^~ u׻l6t/h&q ٱ;,P&hQYڒF)ki*JZD&Q*hBggveq:+Kkԩ](%..ƩC6R_Dh"JPe+0!ULBпA"O<ڒ@0W4g>v>h!e!hixՖ<8\WrxMoO,{pwuw'~T(0\9PyaQ]RET[4(I$0xDZ:T&gX tEh6H]2j5Qb8'PdO)qv<;9'evk٦hB!Ld[a ,K!ڙH)q}fD@䙉1**D8<y,ّ/΢~$./̹#$P,q$' VY'4&ք49s a8ǰEHWc2vkר˸WY-Y7m NvѤaFvW`exNJ'ε!F JhkD8 "d]x!lyCڀ?d@PF␢I'\4ҧ(0P|BD 䚆RLLI`8ՎJ,Ya`(&\plFEƗ<_kdjA\O~xK6tj?ӯU.o[3쒅@ chF) Ѩ|ITK-)'M FG=O-Wa^{CZ4DX! flZhdD>L{)ALO_>KOր⍑g}VjM 9ՎQ崱z>8L T1/4\XmBì/6OnLf>=sv_Q8d2KҖp v!  'pw&B{qj(KFuJh&$& 9'/1:aT4bir֦l>?F~Xz&c7?bv= q$U4qQ_$߾-SE,>1T|pra`2H\ r\77{1rɼ&puB\7ǣ=3jts=T֚\4LZ$j #NXFrS6b]W5o~#sOAxiS$<}#&hZuɶokT96NU徆4OojW1D$\y 2|/껩ϥ=0j\7YV÷ھ<˦qZ'7[Z}+`#^\&ädg+Nǚrv ur!??<8-B]bS?5"esGuo5?aJt rs,#zJhSs@]??|'}pZ@}WJ'1WwA[%t?O t,SS#ɻ:" @9rOW 5( xd-"=/hB.{o Mޮż֭toa==ŬxdY3%5}Cm7Z9b³O0px@*b i*X<:*k, FP&=Lm"i I :4+pxˁY6V)꽁I2!%Y0'nr@p>]Vu\_~*ӍGd'hϢh yc)GNO 0|?i~8im3p=J ؅IcBj/Si@($J`+-ױԚd!`}LKvjӒN⭃d9B-/ #c(K*ĝBDY9dcʈDH˽1 ࢬ9b6.xpR%$!X(OBr`9X鐏1Ugp}J{0X].$})5 Tht<L2T{NH'hA0ƴNE'jD[4 g@>d8.i-D`#ph%7FPdpZ Ob잨2W qS`NJ-'nNl%qm?aE\6'Y!&w5-5tR%(*> BP2 J+c\y+,=DDpIJBr dS!Do:$r`RcPTqT HR8W)4P,4 XxtO{Mf=ۜ4ywl}W.Pd2w*qkq3µJ[7)*$f9`͋"67< | ؃88˰ɕJ8%f6脼 "1a*zj.&nviX31Oiǡm P{`4!Lh!ќQh)GFc,٭Sesv4xg$h!#3hϣF x"3>XB.G:j(;81W^Oǁ9CCD|/Ȭ'Y09ƪPxD6ߜA m[(a{"΃Q Diʃ,eqy#jhI39h*vohw|>dK&WTn{|7Go,mo`r^kVJ?Wջ (}sg"Wɫ ܉d贘3PRm ԎEkN6zR:)x"ip\ʬ&jU<|6,M- @#8Ԇ I:׵ yZпhdn#6F$tuOB&h8B~>>3d܏ErAa;f*)z7 Ag&a S!9(8iOӞւks$ihao'}y?6쯌V=ѭEk`]y]%a,J.xHBiL&UX^Mћ P#%kyg%:łnłmI,"pAbfe̮FVQX CR˒FCԍȲKXdFA=txI x]XQEQjcQpNG~lk/=MzHWRyVdX'").sG% ܘ7]&= NiN hVul}損#tKг` 1Rf5.h$.ybؤ8>X**pX LGd #1fL>Cϐ]T>{_uڹ_%\ Y!WM~P%ۭGm=;tt,tα*'fkom/WȕT j Zwq%1Sh5.t(}4/oZYELeJd$ 9iv*2WJe N*>fϥ̔N!G 8[;FRW&UwepI%=3e1A:g̩KLdPܮ9n?^!n-dսm8_utL*8k͊ןk?nԴ 'RU*n2 APh#; w ' 'mUeKيS.Ts.͔EBfv ހ(I5[S zD YBf;!EJ!L(;m,#gO\;Xu^(C Y+K9[AH<3ƑI+!4(=2n|n]vUi{X8 GLON1Ljd|>U,++vr8R NkKvt9 /P!4J':dଊ܊ e<Л|籍~!0|ߛ/Kx3V/AA<5K>Yu =^v';֬} W~:3t%̞Z ߉~t3$53_r3G(+- ~!ys_pnWpAa3}mvדݦ}W @ͦIp孊椝?P }Fw+T.Zp[BU=ߣ%loq2ś"Wv%ʒ[6"kϮ6%?ۛȝMukGjZKmЧc.;gj]Qzu5k{ZNmv_լ;ܺQϫOw>eڣ絖a<otc5}x~+>YjgYw%k.h[ꖈƪkGqy o"A;n,mG׮eQATaC1;9*[P{BҡrIHm/ptEo&KM aDcr\yJ:y60xFl׷K̶.S?ݠecP?O>s*B8u̹.An Y&./)Tt YldNpz;z~Fx➑g$&yJ܌j'O11>W|~~‘[*2q#)]Z:M]bRkZWzWMd&3UtJyX߁dhPsyx۪~XDڱbMY {\f$R*l:g!zl !U.irC)ԦXgS9jB9 c}B}urɦ~ϑ!u5}jH??JE_]4 E(eeB+B(rfiB{P..uf]rͺVx"Im @0dP)XZs 6 .{/c#Ct[t: $$BR^A Yäen 6Fi)frM=NoF $6fǻ#RMmu YmxչE4awUIָV6墎φ"ITCE;WJ2]YT%Yr!D՝Ci8D@"s؂2pIKTJ#!bY Y+`އf+})їҴ˄:cw_ /Ϸ{=[:t>)wQ@#E*kbQ"\¬ "$\e]>tJձ0R(FR4cSY^SgYS,c2 O[AܧYuB8Yj.ͺW@L238Y%w !|g.˜]֕^81D>܂VIX-% Gu~(f|~Slޯ<65,6LoS#՛ȣ,2Ad'3)Y  #F%㙟L&\cvB+%pf$ǨS̎N5ϳ^쏮?]|+}s~]pa/ۋwC_iK>m$pb2 x4qljho1ju/J䔷ÆqWVhqyHooYU|}Q.[V5^Z|Ȧm~3K۟⼋<2VTśTKy@l!u:}4і"U/8a=ɢ  z#䤤T%+7&ǃi(eV ӓ^yU{lѡqvȢIzB6Hޗ\ǔF8u`JXiwhuiU]cVm5&`K ޷JoտjZ MF%');D,NZÔ9`%c Mަ&x$9qvn9ޤvlmGQ2NB'̅: \)OBgVW,TBWK 8!qUvp2NE\jo*Tj։/P\i!'$`+ɈB頫BmWJpŕ-V~M]ԒB\=ޮ)t,0AkoO~V}\}WL8#τ$Y;0=s:7wKFaμΈ=fvz6YE&r0Lqus9s{hRQU:¼{4/X˘@]RKi`DA!JmM>G#k~yVPmmBA~,]÷_2* WUTd \KD0h_ƽܻUKTO_.^lCJZ֒$U΀:<%źu":I\KNė\wyXt]-TˬRJy+S&E9SJ糩k٭HBx;}|D@c际1^76 Ӥ@R5@DZVTn<y✪<77ucв=62Fp7)orfcṕJWGCWc淴)F9lwSfA)󟇮#^Oyul~}VWAlBO=ytEϠ+Rԩw%h3tЕ%Ŷ? %tut@+~Jƴ{o|.*]!]ybDWL7CW6CW6C+A֒!`gfJ͘AA%#+f7 =axJ{`fAWtSol6DWG_ǟEW6C=ʬtut%٘7DWlM ] \BWӕtu t_Md%4O*;/'?>{彷tCY<|[?8_r#g%=8jl2r_}*˫rs=w`V^vvf᷻ɻSJ/_dGǯҡZ 9._QsBxWC՜PogC '6X4Fu"G{°,,§Uo{/<4Ndr'>H9O?QZqÆ{a]f a!-𥔢滃oǛ;H_#{920QW>~y JpfL3L#w=&ʶd9;CsL'k̮J6 ozbmiޛҍg5cֽYG4顴O;۲HÃia#MZ4a2\cLjb\\X- \RNָDaMq2*%jyY7LFD:\FqZ?-Kbua-6==?Z9[·8Xk"u7ye_xD<,akQ X̘\fh-d+YjHyZ OTE1S&QF9%Q>WgF\{m>OΣ\GuWub0+c. 1TDcU/}g3L3iL)1T!8~'F݀F8LH?ͫƯ/٤JǕlY*`\ Snax(Bx\`eU5ͨ륕1#ϙVҊm3}Nfm)g@xjԱٜÖ-OnadkQ;ŖᩤT~X; &d6SfXKv1`M DL/ UN{9Zj=N`l~վ|k>$04aX)Y=7krK`k "J_k ]05v:'l˰3ҫr L_ *%IB\,Ŋb [L**(:;-1u9|Ԅ ~,+Ďq2HPTH&ʋrbb ],dk5rTPv RKDet((*qt\/AeX.H1&؉qok6{աR-`K ̆ 6aE6!;5G)Y8(=cBi:U`Ro&x,4E'U~%ٕ=[Pm'fH574ZL e C-'P ]p,ʨLhD'0i"n4[SjUZ]A.ŚU kǤ< !.A D-i4(I0Θ \YxnL>O_ [2:l*1Phq ,s4eEp5(H,ΨRX&'x+g * it*TMZ.U^UuA4"fߎCfEXyuABb|R [y޶( ݚ` dw!*Db"Ѡ ҈=>4Vcf#*, "'ʠ]=;F.M7YMhNOcE5#&Z!pB\cP;`vf\殼:!>|ߩCoXXl{&1Yh2X + b|PT8xi=2TdI>St>S Uf&5,:f2X0_*XX芺`FR|&QL&zZ't^װ`TNRrg&D _u5}MANtm>XP ȨU;6\ldoF#AtT2X!O/79>Kf%6Lji& 'Xk td2 ǡA]Xjk>B^s` ѨLl9wKbib#ܬ~򄯁D0ePEy8gW\!n[v#ڱ,|Pt ˨)+Xі20fQ(Zg%'t&X!)k@2Xt9 hU酬iFnrd~ȃ!:88" :<>åUeXJNS cE9aQg4h!0c{;rj-' rtՑp4DF΂Zm'ڀJw[7W3\Ds,e&U$g  |v6.A<>:Kkh..Gy~^^t>ۍ_|L{}vy1qngRX8 C  ׷n3ơ1XC|/D>lɹOKAh;1Z7ySli5‰J'Ή-j =A %f$71`ar׋\&Xנ 8EژZ'c6# #mcaVCJM@2<֪] kP-Y7ɈmxښD<[uA٧K-"޸3 11>ȚAe"UX&; Z&a-5,0z乐??Д n F.a| 'a)Pj⎙$uQ@.`#a՚-Uuvi ZmDЬdW[7Tp/%zvSVZ ]\<qHxT"搂1V/żߋ[//onA9]mG6^[Iq䲵W 'w[9??'!7/&D^~8= ^Wק/_^˶7Z^4曇uz |SN3nN~*g^ܜ0~h L!*L9ΞHU؈$Ʉ$f@@'$1&L$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I Mix@^ml) /q;I v3I MtI ̚:$.QMiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&7 [JpInIl9$̚:$P F@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 IM m) ~CgхIm 0N@ǘJR$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MO[c~= j*W^?nP/vI|}q!\J0v;%a3%cV+e1ޛz/6rK9KTǛ+y+t%hs,iCt%vb7V h=xgCl0fJ ] ed#nŃ{GsP+C͞m+ OG/aY'<?6'1}JoM)~M |  REq/iK#^ۇ {yXkOPNo^οYJ~xVoֶF|ppɼD s &> $ "3HG cYf rYz,IHKy^ax)眛.|t Q?**^kقΨ|I'AқFW7 ?&=`Wq Hbʥ6á:Z GM çCߚQ74|ِhqN䎄gX0Ĵyf T$ OCr2D 4k) {,={(XK6菥$K(b8rP"S!LtGPFRbE?iGli+B )DaPC$pc,R )ʺ*c͑#RVnL}7]GQEoi:x.L)јC \SOc/Cع>u򪤽Yaz2b9z/d;~~gji(;aԇ҆pf'ݞL $Mf!Ba >('ɾu 5:u*#ƅO~zhj^ZxZUsUAFJYJ @ B#8}\[L݉)pc{w'InThJD-6`MHYtH\Rw!]wQ;z\$, qi|;? `Dm N]tUuȦYתY&0*:<~0* ֓WR@g,rA kVuUs+@?~߳7y˻7ﺅ㻷a/Ա)fz~jV)bB f6}T@73gDWQB=bz% Q C&4M)m1Hģ$W=& 6`s5ڢh}쎯I"ڲ -ߕwIֆކZo (r`nis87f57^@Z@=5^iM2nyM>jb w Lܚ+&e:vF6x5'1h&-iT`*aT\ GTJ)歫0 x6MQYl+g.,.!]Ս.t6jP>Jzi.I1 3*,сD%! ೏W_:"?‰CiC=ɲaK.޷U6}ڿ> ֈnq"E8G7jW[g`o[x]GI$-J`MZ(Oџ6LG۹BB,[JL.0"4Ga)i&]&zRW֏Dse7Z!1HDGbq4efa &ȹ2ۍĒE*Lt;.ocd uLR.i#8IuJ< voAX̨3hQ9b0XJBbjCe n Rx4cS҅YgK_omp{_SYf8U_^ nB”]RH3#!K3Ey|&=Hyz3OAY~V{о,(G"i2ɕ ^Y)V>DVK/{vԛ7! j7{J]>S>~5u×G]?@gGp'(bB .,REJB3n-KO¿i% Gp~4111ދ`u?) C DI0w1(pv {Va q ti߀v`ngQyÜ60orZ2ѥP_$xi k<0.^UC5رbZ5wmY.3Sb&T^)(Qw:Q3tJ8P\̦ߍU1CIJMD ?T3guj=ʫ}qu6NT}oc5{rg<@{7_ 7߾?: }դ|Bg!_a1MhuYjL*1ZDE Z’R A)zrHSqVtMT xJ|a;*qag#P.F.婟֕ xo0+" [@ t*>Eyd ,C$st.H٭zO@塣BTN$HY%hxBB]vg >-֫ɂ|(SB$yS{ ?l x1U$2yM9Xy,ǘ,l]vO~GZ!&8[KC$X-J11hE2xPj׷Qu"jƆM-)v5},!] ]vm.[={`3gF܉sft[=t?\->ۚfU@zvuEKhd Ed"5 4Hyo,GONEҽ)wi䋱/S>?N!rFꝤJ6qB"'\ҳI2bp.YAI-L EԾ-wKؒ6}4q*Oo7%'Ji,_N.~aOUo'/?/?^N\nc {gt1?#{cx~Ӻ7V~X=q^.8ؘKFǑy4@"np%ix&cc9κ.TͼY`2NNx-z"ەRGWfz:{P]lp>Fo2 |,i2Eh: /zQ+:I:1&5mDgdLDZmA- 'nKr=b8CBꑢ[g miW{N,{ JZUiɸ2Ք^R#җO3;ֽQHЋGQGĻy%K$L`ACh?tLP6YT(\LpMQɪxJ~Ɓ5ܸN/i>[@Fl5Y eӥHE''59B]A#;,6I/཯JϷf}~_b8m]=aމ"sv='@aTCǒQ)Fu2G w ױ O~e1dy[ϟ;(xL,R>ږm#gX<ɯ\{.:_/\<úmZȂ5@Ϋ[X$`P my; uP"ͻ﷬lXu~y=ދU {mBzmu.[{yT}ܿVVN8$7lRO]ɸ;[ͧrU^v{9O3}<̦-[qwOOKroz.Oċn-NRvh.uSk%lI[n_>|EHطbd=m'aU=NiȻ2Hw^Dd&kD6?lj"٘ \es4[{=}PO*AX^EHV*H3X% WlFFIEmP'&N]gQo3htH e˿ ;AT%d Q5`di\$ۄPT`IcsMzGly6]/"IT(^ڢP!zD.%EJ!Q-ǘ[-\ IXe7(mw!F%[guŮFr$^o 1W-/T8 v7َ7,zȏo~W)|+̟_ayh8.Uk|Effd[&lb3ն]5,y~aai m"~.L!fVc |{[~X#{wyXܰM/]'_K[ezwŸ]WO?m&pEm~~ v£qNDyy;;>~UO7i<4g]5mnT# ,q{9EKҵ/]J׾X)],̖ I0LeJkA%%p.mbHZs_k|$1t-P,JD#X)(:::ۚ{B};xI[/ȼzB49#gznW '"8ZjԜs q@cU bUnU@Xj޾BZgNNW{f {b.~h݉j/$_te+;ա^ M+̂p0tBW=]UFztU 4p% *ZNW%t DJ] p0tU7h*5=+#Y; bN`1XѪv#]BR5VKW]UqW Œt銈֢/9!~܏Yf䧵6hcqe`jLyo{'gd6^n^Y~4R5<,%ײYcV}Fi#ʀ\*;h;MW`G~4Me)]1`R p`TeE w(ҕS]1`'Q*Z}ҍFCWGoWWvVW{}tZy"uJ3c+G/EUL8p -Iw(ҕR J]UN b(th轺(|4ՋЕ ]U8p_4YtU:wbJt 4D.OwtU5p ]UЕzͿ`) *SV[w(i]F`X[*+ *ZUEiHWIj"*(Ϳu;mJ\_*m#GE0-naw0.6`X,$ىw0Z/nǖdE>*ɺfnxxdq&e8sнOfa]>T˱TB !ӛQ>{U Ǫd:ǧhYxUeT/[(*}zZ H/DB}޿b1}]ߎxׂN)q5tnZ55RBUݏj揃ه <}Ei˃ZsF-i³+I 0'eW izB"B~ĕkrTs_ gUe]ƜBF4­{ɾB{Lk)˜V\@l奷0%A0y*)2ř|D],c%s'^Z-8Bcar;fkv| .|^6 @GKPe}`5Ls;.C,GKFkڑ&ꂼX@@uk(0y8ᡍ}B^m3u1F[[|ˡ` >Wq=FrƎ}V ؆:nv~f!,JҚ `jsMgiF霩G#,Jrk9n(7vuz8v+c~> [E*uݼt.ͯh\JW0#FT|&ǫ4!#NT) \s|kpxqtDo6lg,z>|OL܇ğٸZ[46[G2.c(Ղ 0}kM:9 i;_k9& > 6>g+PZ|_n\|ϧ76XM+? ZhUMf+~>wpoov?k*;\ܰn M|3ȭfsDW84kg~00C2Zlל!ucjz{М l_- ֬3ul5ri)d%:灐_22^X NFMR\Ut{/n6|?|: ij1ќSɰHM};+rNXXg]e[ n8v/awx7:~ڹh|(Og12Ë ϫEk=qf樮+ `9~*bS~ ?վX6;M)>ڽ\g{n?: 2$pFJZxIr<*F4:dPumt@4̫`]TN[bJI4'F(ɢ.EF:;},;0l&aol~:D"i!,۠Q^鍊?_=_~|'AM$\8i QoZG-gqO3ˠ A Q>piԚ&# RwϹL*&&eGmiDqr08/ *H?R3R<B$fjqʬa4g69⍆NgE*A/ya 7z(4V) Tb2K oJtbDUJF9K*ԏi 4mv1ݷIBgP<; زt(w`Ϳ0jgE-+so},y}`4W6UykoRxxsM!ϦD^4;ƽ*3ʹv{A1Do1LFFlvFeIg#]]ix{Ubݯ]?{Fp^v]U,eyfg 2žl"K$2ݒ%[meKIg0u]MHoYJH!͸F/Z9py*:H3kv:5n[U )M7@rX2s Ȥ*Gn2L\f/"i&1r1+4Np7_)q[ZOxx1&_]=R! y~9`;]/tI#VaUwx=/YF.iZ/d@%w^T,WO:K۟|M%\tsUJWg"Vofh0T  %3٫2Y#*0ةyEYF͒,`{n1i_h=GQgK>M\4~3"PL5 3& x;P\` I9Mi_s./7a]C=iPJs$M_"+%vM*c3oV>""ϸ!gevqv.llNJ,l^[|7%' 6$w &m$Ws% -0k1.Jђsv:8∁ӧ:Bu>}\ZD+U ~BnTdX%|>GES.h%Ѧ [SNiݚ8~dd6.Xɱ0G =ӎ%Eڏ{+U0 #IpX?1pyˉ#8}Ƹt.ɸ:-,s3)CT$6+sNu*o38Owf']^7ԝI!ɳЮZA.<B-cG.T*֣]hBJ|: KkuEOMU+KU2 }a) T:z ;4SvX@CvPʉB6;65eԘ\|?tUc?q[\2ZĂ 0bE5O%R]fMr|t#-h%ˠ#./qLXs:8tcZbzK9)$κH3.*+mfJ:*b5f\m1SXݕCxwu Vx902wFnM%5zvֺXZ9R78"t}$]w%ʃpLh)Y5Ɩ|<{gc,YâA@<'mF>6Azmi72 1J2<DaZ@-B'\fdx2";]7Og<"8WQ[?Vj>8-r!6;9nc;7XΙi4 %2@ HY& bm(-ã&D셷moAv}|3ۯ@v ֚lo{nJnYWM,RU1*ϣ;b4 hyJ*k+DQPd!ɣ@$%`/PG1$3Qi'-Ƀw09hܩ6_% ]ioZ,G}RgU?}nK~É 'o1֍|!SgC6]ŇcRX"L槉RH('qy Gg~xg@Sɥr@>pCɺcV0:M*853t&۞Ab'o|}~=~l Vh=3;/=sZ2 ,av"Ǡtd>HK Q%bۂ_2Ŝ`KZ!!gݞ1{%=Yҙd18ll099D18X;#a hg}r]Jk̋H3Wu$,S Bt!fp#G##1ON_mԓ,=5oV̻3{HVY/5Zkb xouJEG~A0mñf?{!ߝ> -A_jћ}㭧R,4>?<(g`RY3CІ=BEGU_髐vR /߭CO10ÝQ%oKB`49!2eDւc)gsIGk>%,p85nh;v&oVϓ|(hpmBd~svdTNdQ#g!kftxQ<2g5ϷyLT0i%ɚ9i} Oʉĵ3"E𒔏QkN4O1]ӀU>8_Xs֜d)+U<s̃B(I\ HM;D=uݴ"?01QO2aƲO #{ %^tM48rMPBH-SBE'$C WLk<;#m 9-'P (nѡO' AdQţ:gF~NP3˲봤w&f$8$d?pi(;Ja>i0[9;x3\_7g0|6ཀྵ%' Cu19o@tv]hN݁)Ow1Q<I.8x2z4/j^(x F{@8Wnmx+7s'SۆDreNj/oHѶ7[$M8M݂j6,8]L](q [RͧUquN9 -\Pߤy ;{sM  ?Sv7 .\kM0ֶo$56]ta H.DǼMZ=& 7A]k$kvYpzbgGzyDA[0;sZٗ<]f{XP&uV_wP:y+w?<:6=0={,)70V~8^mY{ `tO|#eְCuh$CO[@R2jTyԈ5!Xd*QuKѨږ._6AaۂcCD(CMc;D|)ȌN9UG oA&u"M*DHEE-;c"x£etT cF%v"zݖpw[t{Q| ޗ!aw۸5wWW?}7? G E}]>۶hv&u;VdNY<(|P M_Tf5GmJL*RR*J}cj? W/ԩ(T'+"SܥJqQE:uj7Z^E de!W m%G]0$Jh^qNy"*%nee98xI)mA.JŬ`Q^b<}嬯.F_nZ㖰SYXzeZ} r|P`|&@ SƩ@bFL+uc8H$x:պ\f}:Aݟ@ (ot3ٮZγGE`to*grٗʙZ^T9STWTU@4=L&{W5/p5pT!\@R4{W]+*eEv=\e*:zpe"ҿd.f;SfUN ch< Uj4ޚggߑ=J =fp\pnBOJ%lR/(o,;~y( {WoX\er(ZiWJ)/ Mi[BܟL6WH&3Lwp0͞Fynz@ӨϤ]=J޲\m$DYo0Kn]ګqfKX|_O#2>j 9&1\ERU(U}>$njBqm&dٶbgk,p뻟~<^0g%}}Q_HxŘS;J>xG~W1L& DC%ܽ[-b[s!3_G"L5zw)`.q_oۀKRգTqA ( `enLFP*E{J$Je.*! 0A5[1YD֋x߭R-& }rBwu`c.! y_^o_%nqp7gqU@e]lad|reYaIvjROqoTˏl" eڈ|u{uQ=/os_@ž,yg&4UƃKRvKQv=mKhv-"Inw4Ӻn0^ۘߣ2=pd 4%-8qKO!c\k\VH^ (+2O"\.(gry|_|͹0򾐋5Uϙ.:g0.D}?3Ht;e'f]թcį˞#zH{+Ç2yB}pzȎw'o~s_>|S'~p&޸M$P5 ڎ_ oRCx9 JzӒSn5b߬bo߮nQG>l5Q1aWlEd?{DM?W"-U"Cl4Es֕yƸFztDw{+^0TQygh2p$V[u%c "@8ayU-=4Hr[\D~Kh 1PKN9w@!V9"X)J`XwS\Uixb,6>I/R-U] Q|47ƋѰ|T%w\ B 14V8DŽ"w$  $[WiUbOK]{&ǽJnyooWR.E:x5>'wWywn䃘zlhw}'gW8ܭ9Th#E0xH9&=,7:M 9[%x=9;vx#5UfոQh;LR7҉oφu')M"UU16τН5M%ߕɲzD[- hZxZ@'0DB-ME-wmH40m%@p؝.n04Eھ(W,g_v,)jQ(40j]d=|X߂1!+=2{nI)kmwpv7 #@>VY^jy+ƇzVUƗܠ ~wW7}~:8e?BOOppɼ,{@׆<7-b;2bZXFOƴ|#vm2מ3W#dVH>lJv!;مxdJt9R`<4&Fl|ڤӲg% tL(dDf9ZHΔt%S)"Βԅ.s38 !V۞$:?v4~یL}.dY,D[4JJIBOEIB2Tbm)gK&vke8}8?7๲l-foa|.Zd9LyNR*`h.V.rW emK58$ 4+"xE2>db@ *A6jgPAe@ rD\ $H`(Gv"B[żeOU|O O fyQ駁T+kϙܴq#<bpS~aOO82J$q&-֥ZFy'=jgXAJĈOjM 9՞Q嵱>)&d"U,H/ V;A7ʺM-ztqdPDJ[vW0O_*YFw,˯%`bLhb |;M<’P7U.y[)o7NF3 ן.\_:zkqu8[.]7_d8چEuu@О'mٮc8'8$jOr`s)hN*0Yźfn׽O~Us 0mٸ{0bVߌ}]Jr5*6F .ud{}*0 ޭfB3 s] /VMfhe.6+{0=r[3Q9.:7 7,{nlËdSnr_y_ucMoawjTnY9dqM!|*OlۍFd#xL}|k[㸞e^ղ9rl34l[Mf\A}]OmΊXI돭|p׿,pnN/޶?l~qu+h}l 1H9 B9˥[EZ{o=C6vVCajz]IKΌT>Y-]:-99USu?`Cm }W${nٷ= (J$D^o?@XZ̭`#~ gn51]d~MTf6d'gkv۳Yojj}X+?&o^-?7 _E/a6,'<~FV)VV&EL1gY vjh^(Ikƞ{YMtP+Y yBEbFaU 1EJsl &PP8maԶ#j o&h!ќprSB1؀Ib[iT̐ " EfB8h]FuԮtbƨ_e` "G"bX y[ADfq̹8#hz%NAd4ж O@=BD=(H;QZmA`)**y4!%̈́NS e7g#b1q#eġq!:͒cqE:∋F CmZؔ$%g!iA kdD as$Q-8ٞKmȃ/b n(*` 攳6+S[7?P5ʤX㊣k!1pUF?28a%^|p౲G3yܜ =<gF 6hckOShO! &fpQ ZS&c0F: .X 2i5wE-8OW\b;\9 !,Uqq7&7+]>oWǸiR΀'udPICr'g" )C1X)-9eV;g-!4BmܠɦI&.0`L>TiFnfJY7,LJ0#nrqY"Jt43 9$U92vqTS( y~erv|.!D0ުd 1!C4 IJR8҉`p9mԸsWMI)E&9iזpr_nfhr'<s/chqi ,hIuN"S NZ:Ü 'S{O>jRnͥEuwn͂M/\Shox*=|Ǽxq%hRq45h ƒFB>jE p"eVz8o|/Z<ۉ޷ݦX=$ 4|9'eVKl;hU%78]OOgwzeWagC[V|e#I5M[O RTڳ6LiyGJ23lqJ|:K}=qXkG,TAD*SmPsAyTNf_W'}уKb "鋁恅dc,U1(%50yPSkmH !܇;$c!vwI|c K$ARs߯z|H"iLzWUDmIY`GPI@.-L%ʬbN0p5@s؎h{](&?->ɧ9u1ܚbwN%( *D 4A|DPNR (MRr2.$r#eO'Y9 u]rJ}*At*A;MwNnmg]pMU i:i )/]p$3ƒqWL Wg zu};Y#>K*[QU Cgz0\G^yN{gr{aB[p1 U.<|g=zϏ=?< ﴣ[GuLYu&2Y.9`¹(=BrIyJ 7F>)AS>csy+i")Y; Zćev}WrA]('{?tp%\Ǔ!.:8ެw3ϦYrG9y^e[:\N] @JiP8FIaIDc ktEh;3~VwTC^RGR<$lcɸDe0 h9ǙGLqxRZIA_P8SֆPpІc$*qRI.P3ϿHp9\!$#[;U8e̡ޔ* ҃F 1M%dm,fN@% Y\,A;y܁ɖbMe~| 'x?8*i33uHD gʷ> ~I*xzgQW '|s; {k%ƅ\"s|' {d=nPU>țYubOWH7 :=*J)d)r-#5|/5KVtBWBD ι/."g^u)&kf gvF d2Q1X\9"23!N+zADPw6v_P41Dr- '4ԩ2k!Szq-hJq _q)x2ϷoMB~q1$HL*$ P$P$WnRcסy}:"*yBRI9uUTU֊cWWH%]/թ+ԋW]]=`ŞY]=\iԚgBWO#SW Jjש&OH]QF'\m詨| cWWH%[իT_Lt[Ԗ2[͒೜b:2oՏԎ+O֤` 61}AR(UCt!&ZW0=q{feKkr3RrE1ghbLhb*9My,b%$DkM !?ޥ:T{#ny菳0{U[e7S7§S,) +Kf,sGǬ_'{/x߷alGDl=XFL?cHBP $&Bu"LH3!˄D(3!=|^m' h!5il JR07rb]AU.% pփ`jz)pJs*7tE'=ﶓ︾ Rr }'ޏPycL<,1!\Iل = !Șn: 6QyIKng}{9$":Xd GT ѣR\!N`!gPt*g$5ʽ"#\B֖̀p%i1 EF)8;RJo@FǒtXLD 5Frvv6V&'.*0kqQ;.EBO?3@q9'S h@E(s2`F$ %P3W>p4ѻXJZG85<'h B^Hmo0X[wW;*o&{BO-O#ەɾWzRK~J1]B6t&:`("GtJ@p=6M"x|HDH<"HϹX396B<$8O)HswdCJ +\IKu4RQQlc{ qȮҡ;8zؤ(,?% z>lxlb/sŹ\}qP_/KJ_/sŹ\Ź\/ȈQ+gs2󌮌YSvYu.%˟9\gx(Y+Km.T֖]+mNnKFfc2ЭuAq 7`js.ǟ[Ut5f!,eQsnzz^6~E+-7t<_swwv9mxvK^}וm|+.`fMmRXןtcZ;mU?L1LKqO ֺk([ԻPL/qGC$PE2RQjp,qp>+vi C4pϺLf5M-A,g ]eo+slxGEC!ZikC(reiC{aW.]<?\ \<( Z+R$2壅!G:r.:5Zt }ᑗ4uMAH*PCd΀` CȹCZV.gٻ֟fk;G`Mj7:,clz."v*i* Jn \X~!"1&kEvܫZ2!Dc!i8DrнCĦ"FV!K %*fBHXVr5sNDBSw})MLPYu|zᵬ<.hZ4c #J$ǔq٠Le?(zS*_-73), FaM"ES`M3>,gKT KFVMAܗcAn6޼OYT p($W^'B.}@-\29]C.鿗8r6^mL##w(Έ<X)V!z3Rt7fϐІ*O T"Uk:P{qLѦ,Mh}gm%cHIR|*\ ؚoQKiUM$$-m?rQa9ϋaZ<԰8\2_Sgyg֑G+Λ62N#ld3-dw  #ѩ5Rr`:MtFBQz~n@Ji̺\E-ŏ4 x2O{RfQ 7i 4#*F}c:~>V(ȾX|+P-#M!gQUV0Zvm⌖49RQMҧU{y1L[{Usۋl 1q~<c+^3T|1k&U)C9(ș%xLy30)@kCd_:O?޻;_G9N/cVaUﺳ&^EQyssޅNH:;5f25TGP8ߍ{v*q{^+U7w[gWRO}\Whg$ZJC+S 󹴽@᪽)>ٕ7cZKZ E1AHK'`CBd,v1K'>|ޤqF;gch{f?*zUO7(Ě7.$mk'YR/1b}C9z}¦8bl>P<#WG@I2jm LfXI lc#FY6HvF0:C BeBzhi]K60L&{\Ȓ \Z MI`Oַu&noߣBg|+涙gzܞ s0vp:!+G'v(gaLjCK~Ԓ.kostBAq𠜾rWn_|ÿ>[%"#Oa/VҨ)XZ$15hਪVAn{s}wh(O= aΩwN9i u /<זѕ-έKi~ 9N`2*+ROtYKf25/~#`"P xA#vN/cbT^Lu{RR^۟O3⋆2R -QDggrBUG6=oԳo45UU$/έlC|XK G#ϗz^j~ѬqU3/NE8[7d4]T5Xm.ڔAmR9QE-Ҿ |2ZyWl ~f >ZJj0$2FվC[;w=]-~JaުnvxCG'#^_l{pO!ivwHk=Z_N>ZI۵c*P΄؃-XucSV7y j(dQr 8B(2hØ9tRmQ$E M@Bymlj^{e., ^0<`` |l_}%@Bd8:攳 x0`{6b]N dzYǭZ#;폙;oǯ"tB'(NDKV[͙2[v6^t#xƓg%&yhJ;f|NVـ!}ə$$n2(>FUTDqX A@D5qg乾Lj˙0ApGA9" RD27"ꠝT:;,Dp0L0*w]n#@Qך N ]X /u! /t4Ÿ]4BR!r#@bIE :4Lf-Iv)bN}ʉ{T= [q"6WeI Kٕ[c"LpmCW1`!TŝE2?V䫱i<:,.?_[ۡdK緯x=Y_׷enHk'GS{f1!yAJX]d7y7]:ՅO )z‘Z]ټ/o)s{lX,D-z~u¿zpic tSw߹^lVC>yEy,)o8_pCg3WS6-K+!gS]WX4 /vk^SyE}\`3 OEIS>J*u[ȯYf|J*4-0almWR6JB)RGl>.y( [L0tXOyS80}cjcrao<RkzT< %3ut\vx ٵN]&χ/YhwrNCXxeCE1p{KZ9$Pb i?TxVJKI5 f⸴z)"QZ|S@RS3dp63tZe2.2B O>rEW#Ü˓./ ݍ`O9>Z7KIR%leZȲV=7)dౠ,_Gi7%ГX$%cMM9{J PA7*P)kSwoJ2͌NOʵsWX{2ؖĬ}PY|"IhFږaG6gdnQїa1DS[6 u~^xǜiB e Yt5M5=#8 b*.UFR]9N]c{28pH׳,:1 #Nf[2jd$#>F4!LgDd5i$[mmskr P2"dR qudA͍ۉ)# Ū(IT,h' 0B` j)isŧŝɜmL̇f;>| ۰(jkG3Oz\NX"yagh2|9qzUoWP} $io5 j߯_#hh#{8?>>.'lzZ >R+z^dx, l7\}/^.b5 FpːTaqDCWޜx8~20Fl/ֵ pRe7k;\C_=/Wg~vuGΖ<98;Y,%0fugѬo#ޫ~\ոZd~rk?,33{wv3xlGtu|ͯV5 I3 Qm:2U*nHOՎDȤtsj75;]B\qPbaKu|賺WϨ:ʤmvPjO*M{43P[Ծ vׇ:JkCmp[g#ꀝph;骣>s` ]M^<`iv3Y|vN^}sMMJrP3Ѓ3^:Rٿ*3^͵ki`}+K0F_֋=iF;\7p;Zgv;Jo4 iZH)٧&Y ]1/tRu(! ~t=X' p[K]StQCǡ+w_*r^NWgI;вL>qEW#jۮ$BGtpI:ZNWe]=C2͹O7th:u}<]V@Wϐ#Uzo { j骣C2 >7tENڛd5tQZ:3+jmu~p(h:]u9շފ\I&lѭ_j엣`dŒք2V@:ϛ.]oy_xۻ~UŶworgx8?= rƫ woW>~tU#w1yyŚ`EwփJopKgAnn,6k=[, S\]VqGT+%1Pݵ$FrTT +7Ena;q#nMnj:s tN!2 9/_eRѐ4ZUŖt1I[N]:Z)O^4? 9[hg}EWV?ӳIR-T͗@UZb&V)DgMa}הtTΒIyszHKMcVŢی3cJbMCSƪ}KX|9DKrSb;Ѣiņ`c 9}{qI5fK9K2ssJkTLs JUદXj!'aZK}-mP31^Ú1>mlʦ3)SVUB)F幰aEk3IEu. !t*ѫk!mNY2d ChtJ ]QERhx&rh=ٻFIm~9[lTE>b( 8Yh1FPTPok5!pBsv1qಲhf/9]炗Y_2ζ<]`Ƣm< bzPT8xi Pӆ4d飒Ql.G_D B˥g20w: `xc$_4)X v=!=%EHiDH2 ! /C$#]K>xOE"L82/*:Z+P6G͐ C@8rš"I,TG7"?bjd<&K΄f(Nިw5xnW?;AUEǒswdBF1cc.5;hː9H"P)#rH$ @aZ55'^ )ڽ,=ip 6-Ah'R]; !zl@ IvKK]IpjFz,m0Z-C΃bНH,Uc<t&% I@i9DHdy7lD,^Ȉ:EadR&tSQ!}L0]5,j !DҡX}q VM{>B:iˢ9k4Q1FF؅n.?{Ʊd /w^dc7m\`7O0E*$pV)q Gi(g8]]]uOuWTH=o!<  a$RL/9BkpZ ) C2(kS]X wy!8aүҗtU/NaZ\U"'9LP8tY2k;CFƂnK>Px#Kp02AZ3玂NFގb3cTv AOGCZ0ߴSac6RSr8%xdUetG 34EuK5tR .I<ց T /-1'0H9+a%D}z bU`"O+Y0p"6f* NN \Rm$NFXX c%, cX 8#6L [PKaZ@P?0 !(`-I{1sr-l̵pU>:]jl& >wD&%!Mx!ܵ.xN:.D/ Fe*QUpD4JEVEeTlWzG˽~40\| T#, \U)BJ+VeW %jXn R.'N(c=3\s=(U,Dؘa]<EZEYȳ~L i!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 2N\61@\a TV3@Jc @3d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2^.Pևp@ .i{&H)2^"4B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d\&P& Zp0L".* L"TzLrd!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2^zZeiyz{}Z\?PwH)N}0Nw P0KTnJ~0%V۽'.V"q%,9k+.l-jz[W0L:{;]&'iB3` AO?woOGV@jhZ~VT/.)@Jb]ޛmW~QR{oяsrT\&z5G k``>FCpO?,ØOFt<kQJe"@ՇQ}a^,݇!~/^<,e]T ^g/x/^, Yicft p|{Kmeaf;qN( ~A$n*Q+ ʱUf%悈R'B*{q5 Ut ˥IO!F=BOn+jj4x|Yb$VVKQcli58#[fͤJ8*b z .#{nYJ(*yJr{0Vaѥ/,1Ks̍7)|UN)UU7}*'rU?.ve8`$͕ FUޗݠegK^| ~ZގޔA=餭*!< Rpn z ^ fO j0x3YX8oWm<{_W-<\cׇ9e٤ Ekƪ^\[Iq'J#1/c;2Lkg,l.e_KJ%64?rR}fq^qAiaYY8˫fxO &͌\_!s)mɲW!x/ީn4\]J۫uQ>M9^\iVge.fՆ'WGc i<͏;5>ڟɍ@`cr욉Bw6滎+ PL5hs81`j<J^(K3G15^\ϲ*sD:tTs/oٗnbі\ŻozhK,VJg+\Llv<<; FwW/\8WU5͸ػP|}ZUbɮ6dKn*N_[̧Ўa1o+1?o+YÍTKt, XmMF/z])PN}1x2X!)NAg{:s7dSά˃zh3{jjݸq͏*ZI0]8c`>?|q5sV #6ԍAs]CX`CXb)/pf _ӜErYWTnf<+1RCQK'U3W_yYE2!uUзUkvJ fݯ3Kꄰ&ɔDR7ƫ譑9egĖ*ehQFssw4w bޖP+45j. ÷;[l ٮ;:V1R{HM4WCW׊auOBPYEgsِR\k 5c=Lz'.2 q1jIxr]VB3sԌq%HRS!%1rqN\;ps5VŽ$PKox -cUޤzǗKw?^}}8n'+p:G="w6r ,8@aSDؘb&ey*+>>^F195,o{!CcPe#X&Rp&~zZEgFR'-GKF'D\fIe&R,_hv)jM}kς)Jr.厝(-И":eΜ>0)AS7 _LC#!|WAlR ͹ZTzy*4N`IzgrS]T҅'%0fa~ιZ7ijFm<^ßLx9(yRαWN`$e;$GJ):ÑL҅|=L^f7$'x9z!ɻT K6iN7~釳Y!Ei8T_?F4b.6cڻt!Ofϫ-ɒ41r<(.hrMV ߻~q |94>[BZs~[{ݽ^_4͕_ NWa~p_<8oe߮$H@Ն]M볕Opѹ4I֒- fX{36,o`"46 zb뼣'6m/*#:ɶV[PNJF;HS}Ѹ&_M~^#5)v ʈ t] sPǿǛOWgz׳W?9̜ٛ__4p Gm"P&AMoo񮚆Ms#Ӵg~v9[}ݥ~Xzq՗W$]GYlYz͆2rpuŢXS=kR"RpkbClvs5 s-Ƹ5FZG'dr :9Qp.<I egȌqS'&gϷ}L:9":"w.He"80m4)O3NJctW4ti[byp_q#TPzoC'V{=W6ؼ~>[E6 !\k0w ~䛛qhװMgH뙐GCZkr_qve2HW3W)J%+#]xb⩋]'+va4.Hg ;GT8ufWtPRF6묹[$"sb:ei,0RLj!9S1 OH1p.t]3q M%᧖"OUy1OY,sSio|Թ MvuK)V9-Rrsfk!* &63C&D;CoܝԫǛ~;6C1x5k(v-ـo1y "HMN6A42ȹ(FsɵrJ(k+GXYRˢ&x"cH&D2w'=:Lʼndd 3`GO< J;Q$ 1G#ϝbė[,tfX''Uȋljm{} uy !d']|]ŇaSd " 2OsTr!|$)~r DTr)rIK}Ƈlu9`-#:M*kfm1m<}O=}|9 qcvQ7+w5=߳H@>%N>="$u9 6 $ڪr˅KCRL*7u+/'Kcː J#{N0=tFTB@\GMWAi[Wya^t&;q ޑ$zm6BғR|e &a ~gzWta{=k+Fϳџ;70Ơ~H:|֒{Dg\Pd@t6a$0&iQ;G]#61Jyv~7.Bodӷˠ"~za4&GMLg ̦ˬV:)$PчTB990-Y,BGF03 rS2\( @t"-pȨ%4Vd\nj9C U-B x|%L=$$F=[k;ngz$/s "|1$ O)8Ԛњd4g: ma4ym1A;-Ǜ23)3a Ap:ã __ RDrR"7"꠭TlOgWAVҷzZp(`PE=pX@jXBVK5xQu]Ѕf9pqC@yoI!r#=@fIQ0ҰdyV9G)`U)'xҕ!ħTŶiԘ|h;\LY7} ~K m'.Oqw,jAkԇl ;ˋO]Jn~SV} 27Sbʯ*=tWf}X>J4)eKٺ ']eɓBvi]Vf;]M~hYm}:u#ټŷ+1&m_&yͦ+B;D\/ %6(EiZLsLo(0b`/#܉D3KZ(/YE:m'@%iJ\yuuǧ$b+Y q[?`릜PRU*7'Bmebl7+ fJ>·ؗ;WǤ9uo&/4$@!`ό{͑{fI`"D=֝և{f>9ߓn&޳Eu3gMm* кM(?I)Dz6,ibmo{nq7lxܩ !j?Ω]"3&[+0'33:Ԓ{4<8 " Vr Qv n_+H/Z $FlW18Cz[rT Eg  ]&d*i N1ze| 4zqbi݃j5J_ZV @!RxO^Aj\ܻoO^?m]t֞L!3ktY ɱ(OzZs3ZД/˰AHNDD.M<2T$gFu7*RYi ӃF!e k9+Lɩ\+Hն9laj8ʶЎ4mA=abɻvmפO~נz-R 9nCI09GG^V6!`fD"X2Y&)i[Cʬh Mi J \LLEʡr&G,>J9l|v\̡XjT*[m7Z*e )e4de&Hӛ@K^`vR6Zgr{`5W!2䊼($d6dy,Sk&a{:,bb͏-:駚E<5h?EE2 +eR} 6mc*gb䁌T;37F$a\&P";RVAA+W"$me98!BGQLB1 0Hϭܓ)KRm"dKK-0 v@]HHlRm\M#t@krҡ"ulhqXj/}(+v0wrui@a-OYxBe!"-\nL!:ޢR<=cSN`wҒf͜gM&k&N&s}MO,BJdy )e^TJЀ[QT( '`,PrU.ܷaJ3,R&a31fcN !$JrC27ѡJx#o# YC.xrn|֚ '@w!-~\N={3c`sŹs'i(@BSX%YjhDkz_g=YyrL}Ok䁜1^uZysu Ӷju/D=H8R2> nn~ (W $h>PG\ƹe :H%aSv/08; ;J[Fs^H$d`4|Creڂe4RrY,c-u-!1c+~zuJʾ\*p01ydF#=kezPg%'6k':i93W+ءP¬Q?zU=ڀ74Ho}W]ACW,r?h *B/W5$xJmb h}5S# 踞})7.wVx$jt UZxLI8c:Z<o2h,˄pKy+s=4HCtj THKZVXz]ƍfSzIH_?j`;v㈎^ Y#::eb9GClՁMT5铹tFUTd :& ]Ɋ3ވ~DzyU-dPx4 a$G^eRq%t:gU*Ǹ9 21;3(zgKפ$;{}}WB=bͽCX.;b+A=ܙժmIuoAF! @t<#;C]=bgg=z/t^.*x )4=L$is*{ޣP'UOR$%r9 BY!l匵,NG& spPzdqtc7/MwG oW&my~sMNg9M.n"nƒ©+JeqzĶtq{s ?B%&OxYY[U*_S(O3i3x{ٓ-@@ No'\U`AnQ$d j' 뼇,6b\ohpl.ho!B?c.]ݻ׺5}>gmx_gE孊>>ezdfdu=ʸ;!궧_ܚÅ?YR5xl8p9nn-tmsu?mԲ{}]Ý+1=?6W;9T]{@Bu S0No˕_~pzӯcRZ H\63F^'FV=FڜڧZ("7Y{(nCC&!3YOgvHH:Iт.iJ)RTvcLs:0NAIo?  TmƢsB@8UPހursD \C"rUh}$y!Q}ʢ,=]b%/s,!;VJĤu0\dĈ-䨹4wsQ.0{T,\yf"X,F!175XN<҉"R8=YHc&2r 0p"cҐpY[l$5&dg\{#W5%u0g:_QO[FJlsouT3GūapE%뒜M3^@7/5Ts2 V+KAck7ֺ2A+tmE=*X/[ Kݵ L޿OM#pS /AZ:EXlZYc@]QlK )7]kCB,8i-:k*(]hkckVW?tOȳ8vu{26jmGt<6W )}i3׍Ta6ި6tJlwh0ZZۦAwWmA[`E_&k"et 8]WpmUv( _E-kOGv1kUU+5LƠr;a Ɗ;Ľ-qҋi}ibm~!%lrL& dMJt1^?禌̛_⪡h&ᾯ`QWGz˫ʟux)N۷Wg~}o?$(6כ^f/Z#nG)(qĈ'B#ʪjkh;G8GglY$w bw|?h-{롛~P{L/erv or^_?/V/?~%j?szpbVLwW%hB|rKˢ=^ak\آ^^]G/_m?z!7n._=Dt|w٦*ԥTkLt. :H.&Agsu;úRtmGĕuյWNjtcVj;鰺]hJ#E-vu(q4c:* 5mYHpW'yzX,N˥,uko/AJmJ^Ɖb-)e*M'Uk٦>j$tzJ%isuhiR::_v:m]eUhKQy@Y;Y)Ce[l%O 6-_ouUznE W]YKJUأioxKj>ו%ۦ접<فҢܹ_c%z|*|Z{ֵJX[W2,4s&8UڗdXB},pO8~H==xd2(?9hr^"ة^z1Ӌ8ڠyzEDbO/l ^ZRK#]!p+uEArZ+RQzu@])mF"` q\tE*y]VYW ԕN Na+µֻuZd]-QWF*B`'W."ZSQBuFz銀-p i(ͺZ1FYFB`| Wj'ZH^WDT"uL[/47w4bsvAF.&1Cʬ>zG~_.nomjal>T4+)N(9 ^w޹XJMOx67.FZ'#J-io8(>fDDk u]%@u@YtNװy@> ҕYX1p̏h\fq$t"t岮vz KHWD`+=#tEf]-QW4b`\tE`SQZu@]iA(F"`GWk](ȺZq2N mp+2GW8e+FWkpѦ?N:GWKԕh~/޾8F !`#Xv~hw7W?goh#ua1r ;YE()ڦ.EEJG?t~wkld'pwɸiyޟ^mǴ}eyk_kjg}ٞwl +UvPDdhv-uգ}~VcBk9ic}K׏.^cGoZ7D]d\G' ۶ooH~7ESt|׻́$UlDkTQ ?1hpBu@uWj|ݍz~F׃ C;V4sXK ^K׎Ys/7//W⦨duQ/WG~hmUpW[!j}RvN, 'ۻ:#I >27z{Y^p_OlߺUͽc@'o|9y ٫ 05Bۥ랽i[]NA'`"M\S$.PF{FF%. žۛ.Yȏ'(LqQy #'%6 $(1JPuVMQN7ar,.lɊ[D*Vxc^=J=Dخ鎚ߘwԔ?7N7G:KSe2+tM=5NO>ZP9gwmcbUقc/AO7yԃ9:yҜ"d OKNsI%ڐ*"/ZbB²泓&5QuֆStEOtEu52RD ʏz,<"J+NA|+>u]sWKԕNAА8ܹ А8JJrS_ {bĉcЦ4턖04"w*M.M k(͚^8FWk &(}[ FB`0|8$pK^WHi˺z1 #> )S7Vj*̼V2f QZBBծU/13{FW$]!-r0u])+ .BZ+T"JȺZ$#]ltEqڹc(ͺZ2 *6sWD}"J箖+P Wy."ZH~0H-κ:` #]!T;ҺQ湫e*h=Q岃)wgشov^D%g.^%0r`n"];L?+bbh;Ɏ:odZ^r-?>>xc[Nاffw1muU o\tƖeQinVխd88>АюӃ6sfulb6$tDZXԂptE3JW JqѦ(ɺZZFB` )Aꬫ+ϓr^]j5"isXJ>nW .S D τe~μȩ/ XYFWJEJe]ZRj9 1ltEVqzR u@]) 6-〃f+uvqYW Mt6"\' i0(ͺZ*FB`hp^#E "(ϺZ@M+ lHUSDiTueW ultEApѕ~j'JL6?V}5t<lvZI<q|8%?gGeM/P^kКx}Mp\gGQZKU8tE*(u^ȪBt<{Y캊U3QttծU/v FW;wzl\[QD4 R]plt栫%εEejYWѕ6HƱZ+ iȺZ H'#]Kp`hM]WDu`R8 = x7u]\t^uE:j4Վ]!l殐6X,fOYW %z*iG|zb>"œ/pSď+lNFQQ< W2jYZNx*n asN#ևV/7qZ@`|fB L-ԧr03.I_WPUlb %J粮+|QZHWROZ Ѻ3eYW/FWfd՛ca orKpU(}bi1&BW&jתNg#ٻ6,+D&U% 8{0 &aT {̖He#ƈlUn{n=\u[*=}(>+oi WǺ`?LWORU|{C{iC噮 ]q B2\uhcO>2JtJhA k]!`CWiB<]ev𙮾 ]I&-+ 2\ޚV<]ez'DWƈ:E#ߡR@O0(Ӻ>˭ |CJh W4D:MgLODiFoόh+U[*5'OWyF)ҕ1Z1"BƠApi ]epd8t%vlzB/}=:՚!u.IIO7ߟe ߿MЄ/VAgҿr8u6<-s /G~"| qF "wExf$%̟?;"ͨNǛfxB}uW/=k?sλjbWWv_sQ[; ޝe^*gK}cۓOSF m4|2|T? Q';ڼFez{ ~H_CAB]ԝߒk;V!; eK+N#;,Yg$5z;fe9x {Li*Mo(9eW؈P6"EU=(.0΂t<}d > Te QjA%,BYD,aQѬSm;6abhPo@]R%*MI> E6vo=M.2Uc# uײ\ErwW+A0 Lj>:fmtҼR,R0a-7^٘*PiMRQrA&Yj伄":Q&5.pEphiFPda)OIz*(z}~]߷mmϖxj{@e`UV w|g5 ͡|:g\o%j릀)놈9գU۴+ЭY>Q맸 H1Դ>Vs,z g} 妋r{- !Ox ]"FAQȀׇ܏RyA5 t҃ԏ4JfS51yJ0rL+iJ> ,c3FhL9u,Hə*.v!R#4Z"@$F&bsSO(`Nlz'n`{RR_V^&!yŁ[,0aphnsUσY3{.;ޕ:][+:)l}6{Jo\+ʇm5 ||XoaKEЊ ֭f+kS`[3RP d£f1L3#@b[Lrs>q zAqv)fy;99R#)PGbhkCE9#F4u 9/fnXz0S%c^{)Ai JoІ oP#Ŏ%(.v폅AqAy,6Yձ;.JZF8$a@\z%N$`$Wbl+״SG`2TƤ|A ՒF5ZF@DQN*4VvHvCE!K:񬾕^)91OHF AĨu|*Ϋ8mح?# zaPQAP1PEuYgXSDd؏c~)w.ôﻯO1˯-CSGz<@>Ӽ@c0eއI=;+7MgHW84ųZh9ZVƁDg#A''Jǭp9Y 7X!~q.lEq4+nU;[ő]C.>ƛε-dQKX/[W0z.##) =[2mE E_+BZJջFowZ SR>x|>/>xQg[P%|e Fi8/떙uVHF?g*wGsBrRPsˆgSη5Ǝh\p`ǣͲ<W]36WJuYȺ\k˪AU<}$g,|2Hͧ_G/q9a+&Q62|>_ӏO=Pş_s:EW؃!0_{֔55̚iKֲR/*||4Y|_nw}E@;+Zt u~y> ? ~,ye5m1OQW6v]~D/Pi h )s{-q܈16^qn].CӷYT]ZF!~YV҄a!G<8cN:AJ8aNcSV!Xo e F 1Oi_~ʋ:*eXHd{efxb҇UL$AmN XRXj Ձ1x,X!-7\6VhcD h (#69kZ*Ge-ZZ͌a|_C㫫%U-$2 D@sDT<*6>m;}ؙКQc&aӔCE*O6;kTA cр z%PSd /Pt]"mʃqКh@ Rp.aS!Jj|Ć^% )iIr%RxO)ZeB$b:A~ i_8Z 7L[dmZC4Ί|7^Fu:H)sefwެWyP$% ۘo9ѕx [$f4IR"%R^+:}A& k ڎ:᭶ѭ-5I:CKAMmbAV؟@k/z=Ӝe5xCNI"2e eZB`m#ɮBW0?H@ F}JeRKʞ,y$%JbS]%"շ=S3$"fx3!Oh#)D= ^m..9a|. cM!cV--U6"0<9nv( :^-5 X sHSR.YBy&mS7p@*(-ʃjMė9\ĐGSU91*ޤ8e,&YFB\ZPW{5c o iXxኵ&3W])4aȀg:c vmnk(F%ӊYUIc QIbY"hHD1cy}<̰qtf@m2ӪdpҫhR0 R&V0 ANj+H3Ԃ ccDN W$2+BL种UvkJhDhEYRҎ2#&@܊`mmKnjM,TG''biVp6SRe%)8 dxnöμs${QX] $S} \{#@ |tm^O_) 7Db6y.e}t7@QR"d+6!o%tMކF]Da @pAs^ πB.dsnrXhs^7Ō: hIbA%Y֨48^pӁPE \"N 0,1Z(&RPƀ)E+Y#Ƃy9ye>F/~ )dlgZ~ L#rZ!m:B&0 ֨ `f)U UeV- h_S[ 3]$y6$,2 c`H56ZĩWV'KǴRb!fʒ4YC\tB$0z`2o6:S[LSpRԑv j]0! z碦r8RV   Ly}0wC{of{Xh+NJd&yÄLԃ (0 7P#F' L; vX 6M҈ fV$bʅs <@b&v)"iTf`) ;"=hQ0RX"2⓪`bgisd@Ֆ"Z`p܏!3AE ѳdE-uƒfndjH 4YH>9NCDY,JLT@$4.p-c|zdmנ\U8/hEo:k/5s%F56B FxX@a} 33 S2Z0b"z F`JnÈ'9>X2ܳ"ڑ2_ebUqy2 RLȎYY 4EL DC(l0p9,Xc.\(zBw*;c70&L0@5a^Tō^^*PN nǽdx?y\U 9(X;Pא5-:Q*6:񿰵>t vu+^i~ɘ Z [3%*x3 ցPE#]}v,m`M9VZ"o\>̹_1q|Z}VaJ7BaK/?.F"t7[oa~m>:y^Koh%+R$K"j&uF<֝:Cu)>yL-.`jFn==4Zw-s)LהR]SkJuM)5TהR]SkJuM)5TהR]SkJuM)5TהR]SkJuM)5TהR]SkJuM)5TהR]SkJuM)b5'mH ϟ̕Z?GJ+vtnt1#cXcy 59BaP!0QT܋l7>N?\u\i-=CeZ'÷{!͆eK{L#{JdT`UIˌ 80tix̹[z3R1zЫptKZI={GxPlC[ol}ilޛ{yo6ټ7flޛ{yo6ټ7flޛ{yo6ټ7flޛ{yo6ټ7flޛ{yo6ټ7flޛ{yo66s߼{}HgElxXqrbth b>9\SVXzg#uv'٩\g~BitvF& P%Y V4cBZst$T-i4lf*DOOSy_'2jqk/val?4B7>!|<l'3N>G rld`O^ V*e^m )]ckLt15&Dט]ckLt15&Dט]ckLt15&Dט]ckLt15&Dט]ckLt15&Dט]ckLt15&՘ibثį:x* iE _|;Ç%&& +J<})+JL1g?/fcv6?oGI FjƵ, l S|yJ;|1v㧡>m_p'g.eT>&F{Cp=s}' A{;Gg?ɩ-gRV` *&Z O\_2R$jVnDQ;I9>>'&Mh11g?ZCY5 ā4ad#"txcX/ӻrqcvD cZK]\"x)RZDv^SZb'e~Seyp/b>;5{٦l.g!gu?\`ek_Т+ |r a6,7#g.dy.~Z_l~}M"ū|w7st}rbL7?|%q[i;6ꢷc:n؛wiHTO=xv glp9Fn6X˥痳3+zVϯl]y;Jrn|3/^/d^?\GO˼g87]X50ډL}<.yZ^Kg%CJeJ-j\])Pfy]}I\%fͿ/hR~].ʌEGhd T8aWEݷۋe`of?m}o`Xو"/aTǔAAP7n%-^M}=r_I^~M|>tdR~rݯ9 oM4ewJOoHzLzo]r\7[.A~o15d3U AE5岜w-ɲtҎ*;-5./pp/|+-&9q4Z:$z:L 2"}bBd\fqd\Ua28 u1 z}ֻ'3x*sxs_rW&ko~wn]yWW7∝Ūrm9{Â#Icvy@RjhD4-ElR0^i0@a&%9z<6 !s}Zň̜]_\P{wy1j j_S5F=x"sA+e{VB(wV>)V}B3qס.֝ЃZqFRԥ3.,eý+ x.?NDexjD 9:sZ}rޕF-NV v+ WB϶vsŰ텈̜o^2/9b\\Ho=ʜz8N#8qc*hT, hRD%V.5pl*8b< W#g8R9n7>QkQw)GLsYg 侕šqߒe>/~ª;pgg> o~Y^CÜ/rJA[yi]qʛbpofIUqŚ>Rʏ ۛH.>Er0/[qݝz,)ޖK?rInyFkEC\xʧ ',FNg'/_d^ܿ?fprPK: Q9]4)¾SsyRx$->nT7G קMvW[?0fi~cQvGq5d/)}㬜S#ٟڪPK^;x,/FeXJ-X0+RSow^s \${tZc̏ZL]:~EQbirl?4ZKc0[ v;- 03U0)3wKwnd "{hީ&fBfp2ֆ1ZFaEp>pģ=9>*KMڜ^s6/!$!vBwPuVa"9|I:n41T!%J5'61C,9 &Yus㉠ǗTƄ+sWdN[4OD,J) ~F.E!.oh-HPʪ)-\x$^ Q:^zX:YdA:X?~p4 _Eܱ1pnK]-I|ɡfqcC1>Fۇ۞?;E V:"+xmЀ4V=K^,9ІHTN,'yZJW#¨4x@׀R"̋ie`O /\An-K)PD&wlp J0݃2| I`[I4R q5рz524^Ѐ۩D xd .;ɶ6$@fZޙ5Y. P` Vhy(Y?a,GХ4Lኃ|p=F7R5tIn֤f0ݟ0ҥ+7;G efeo?P[|\ !46oQ G÷Y"6!nLm0K {km7WP?~17MeR9K`T[z baOm~)YeYu\5w,k0> +#e"hXVQI^`^ҌdhY wڨ`>̤/1LyqQҚj<=(man>Z7nrJXU|f(m@ VxXB!yiRParie/FV0 f#=@U PpjπUI e(u,qÝp@ O\ JmcU9-lJȩ.,/i݅ \nu/'U6۩JY "NvʹmYξm;ӥmضEXmuh!/i*lc]vE&߻3vAh ]Ljl*Zd19\7{i!: nmu O/3癍􌚪Iiɕ\}'>~`}+jk=륰 i[6JÿG(n|`V8Wuu>-V<\ oE` 45BZͦع$6+u;-VRH>R(&),_#XB]j;.AGyhaSs9\ںV?+S~>YP.oZ w`oߝMUrb驪vE{>{ ‘Bh'g.I҃}Us}y76޼*NbnRiT~`y oxSoLuK\럤]E)*A×[4?:$(נ~۬H =?DμMt>$[4o 1pCsDI솙MB,^Sp0,u Ѕ,u9{N\s} יxk#-.ByU4qw@ %]&;$n]!)%R'Rjy< ?i WU! rM^r$RΦEƓt^VWBd)PqϵK.%%D48񬊌Y<[z_bbRKXa[sg͠vg/ہw7:_~Že1߻`7d|c.Fǻڵ=5'ahԥgo})_E{_|e5)eC'GŹH)>d֌M;ZKR,H;Fn$\M{P|0Nnfs/.3ۣ<"Z%-[mnrd[?bUzqGe52V@nSFqC2ڕ=$xd{xVhd<.zppLQ?֑cYd_GmX~[UASrYAJRb{T*]:DiMV!VeH-L`8>ﳫI;x(}gz{ؒ{d|"_=a^QytipP8ej= I2t;M]Ү=E7pisu6|^/}֦F}z'1 ofoKnه5udicsicxa2n6iݲ;h qW^EvbǤg%Kvy)E"I׏vݤm_˯>@<['iu8[TWˠ{sŇiBҮw8Vt ?ܶ&yz4ӆn_r/ Y&oTrQkq0/.S no) ho)K{Kw\n R*WwE]in@*J .*d/2U,t}rGHE~rǒpd14( p|7|-=mSxpԎw?N{BF9%hרUJ  aYXJXʣm!S=Z]^P(8LWRJZXW)՞vucvJ_zNti<éOALq/tyEi_ RkNzM懩[GVfѯdm?kGr^8-YN~Xwfwz$^鋖zxf@{c&^K~-?7͞Ćr~Օ_ڝXWjhY\&ƅ4{O2jVP;4hWC&4ָ܇z9~D.C|WUz}ʔazU'\ڞqMWt",\ WzHhpj-T'\ WBs,"\ -8cr H Wu5D\IMm HԱ*:H2j8BX)"  HW(P\!jRP<\gX8H~2H*NH p.& YW$WG3$O\'\ W1n7+XlQZW[qfmٲ*L2FJO?_ͦ5swxM&ÈS ;7~| #4 6H`Ժ}vRdU`?Cܱpr%sԆ"J*+g1P.\\ibU:HNշ+g՛Sf}vk3{Q'Jn*U`֕+pت W$8"\\ HSڇS[p5@\ ' J H-o]J%+pEW$ײXpEj]B%\ W1nb DMf2HjMARTq,W$c]\ \Z)BT:jrfZxXV{6/&u9|}d?SRԧg3:~ʦ+#z\MŸjs is|_k'C #Z}fuܩi6T/fA[iʵekP׷J O_YQKƱ(H_&A!_^Ҩ{\UF;O?<PT/wE ?+栚UB}ښ݅JvQ*˼^ Y; N@f5J+O `Y/*h*P.??賃Ze>ޮjD\i)0 kǺ"& Tvhճ0! 7_L#OrLsJ+CV\Di &L\c4.tLJ0=DL[ I0 VM&Ժ}vR%Aʁ0^Mv Hm_TТ۳)p$"X8ɕ=ptS{uSi\\V=c8_`d4"}vSY"Wm# :\\t,"…+R 2jJ !" v JuEj:竛Z'\= @i."\`! \\0:HK"rLp9Sh+R J'+ɺ"E+x4+wR%\=(%Wrɭx`=`%)SsDuj7-Q fNإ+*5>Q«N,luNLXg?&evbSGBST܁ڡZnPf\bq-Z'UZH'Zc:HJ 7Cĕ3\ꘂQ0H H.DLjua1ҲoW|=;eG/,Xrm]cJdT틫 Wzn;]GW$WXpEjm"N%\ W81YW$-]]*uEjEjUJZ]LBhpErĂ+TkWxqV+줌W` Xpj9q* %\ W:b v >aoG.t\J)+S:"\i'ge䂎Wo]J|WCĕaRme> /~G?dWq-hnvK/`iz v_vnvz۱:۪ͱis|=g  lD vXkh|aVuQG.YkְFW$D #O=W҉J" gj H ~Un'I?%{7ƫ!gjsIśNI$Hʱuc(jD@I@n*itσtwá+u)v*7$CW֩J+x>qte!13y%Y 8ϧ¸U Lr+[$DX%DטThJ"M[eK@v]\%S+BmrݜA!ҕ܊ވ`]\S+D%g^ ]AO1`0r8UZ7a(MdoD! sUϵ2R'DWd ZR+B+xtE(!!ҕV]!]^]\R+B+el-3]턮0N$DW(ҡ+N0Hh tut%PR%DWdJRc H>$dtRFH֌CW+B+Biy+mH:T;*v".{WIWNr&A'q `06qarI<.k &;#X'5Siՙx -ǫ82Hdfw& Ӳ33qzVN1InbmjܖC_[LjdR Vr{jt9pk%W)gFBdKmbe5x_M.S+B+EtE(ux=t%{^3X툂s0vDah]4t%3]=W 6!"+UZUtE(tut w]`++M-PJJbOeЕ ]!ZDtE(ck!j't.%i"ও" b+B3]"]ifDR 6 u=]]$];+!΀ӨcCvJht`7]J2M M[-t`ky*tEh]^%tLd:@rTR9;5.n79]!Zg^ ]WXj ľϛ=ϓ CutRGֆIЕt\գ_C* ]\R+BkMt(3yBtd* ]Zc+BpLW;+|H$CWX*tEh=Nfs"g:@rwEIW 0.HW 8JЕ'R+Bq4DBtEu:u4.v"tr`Ne"H\O 7QG%ps~B,Y5+95A䝱xY [zY*$3fDFtqvhJĻ5@n.TW'xT_>&eG+?gT4$aɤv~hn6Ow@aָ ^8+&ò*f>)=U+!w +&oɸY'KN}Xd:{ZwW#*C^\N/x zN?ٿqƎ3rV]rvWS+ $@{r4k|I}~9MPƠ~x,$p_5=GS)6e)U țg+|ѻ=Fif*J}5. C) :0i t+ⷰy73]/Yq@:lyY  kw+qntϧ+Dg> I'ʖh%coyg֖7pw-oo~Jx8z\?Uc+6>)xƷXŧOn! YdeGS^:o $8i%J`qͫqM-΅$K2ϰ:2%3dpLuoF76ӄ\WM&DZ |/y57J GUвCkB-a%WSjl`9 g# :ܬe#B9MZFk^ 9 I)L0=,{ zlT ՗*O7Eÿz߈ - ґZʪnbq)֬:۬3nE"[qJt*]I䪔:!` 2L$ g 9Sp[Q7Z ?q0 $»W++b#t0bW2-ipl$JE`ɚ۔qFsxImUڠ+nN)4 [b%W27-[iUNLk9[ 4`p}V`q+߼FbS?:jtg%WXaVCD{w 'W>8 -. 앜puvz.5DRP/7ۣ޾Egӳu-փҝ,}zPj8l*]Lf&`>_ 䲦M檮+pv.)"dȋO}]5x;V^`T7լ)ST5Y%7:Ǐm?EKW&f+֌Ujn 0i]=lG)B|㛖>z\Ww$?e"lF29CSчRo1^Ş~})qEx:+nPURnBrXXUA1fdCእ{9à7x8{K8|}C Eٝ o~~'w_SUԷ4Q='w?$5vѝhi\ݎp/-\6t՗R1or꽟^]P' 3!mJ,^tĖt9s7|[O]&ȗVUg.?l9{q}٭>Z9{1M9]y}~OA]u4sz{4]'j'zr zn5mh.sKj2rurd])07qȺrꖊX98vNhcvK 3goλ>Y19~:Emw+:SOfğ?\yKBBc!kY(螳翿C@ѡ }Lϛ3 rVOj] WdA!.khaUT45jbn%5a ZZ^Җ*eK*$%H>bJ.ZZj;FV%k_ sP p(Hjd2shL.Ч88!H\ڬ娧99A)Lݒ⸥ %̰ۓ׬3T_dq=T/LJ#[i:S.R =V\Ci FՐLJ47EjawOP+DU{60BOK] i|B XH)ښ-Vm 6\;WF|2L i]T#Fe5jl0/+;+z#\n}63 T:"L. dӝS1z^' 87娣 uՙ#[5AZv6gR 8@p_er{|sm_ݲ0+ A+e[Nӑl89aZ&Qg:dVg*>+P=RY2"5L0''x @j uhb U>k9d^z NS ŲfFzNʟ `Ysp%v5Pa3"瀫tJ;dc!+9E5Dˊ|d 5H"ǜ=:1S:Åa˜spz,%G ۴ȃ u&Wټic;\efM1 R9dpٳ catvN ~m9z~[98|*'' M9Y)ͰXS8]P#̢4Bja*[=×Ώɥ!Fr4-Z3[VK\aiJ^k7I[39lo*T@6QeY##ɑ>_ , lVm~ԚJ=mMۗfVLjŚ!4Tp4l䍐o.PŠp)F^553rtz@x~bVU̿QܹD]gbUg0«A/US5Vk^ɔxs,ZN,OtVrp\Дە||O1'E&ᑽ0 \25aYRi([BE- Qd9]dR%HםT_,Mdyeamgrlܹ?ѳVOY*l`D<@D|-EV$OzD|x41L*JM l!k6Ā=ўPQw˨s2፯ Y-ILT$Y0/\YJXdJ-Y9ZJ2Ѳ2k98H:{ ն4N,H貙F3Τ2u0H.x|݂9YNi> @BujTXe-Mn#%|gQ35)+ͫlfq.hNn)q ?jg)--_ov{$e<*L!n%p.OW 1St:CW2'Ҭj"B YQH֦TzgeY{ܴ 7= :0cE\Aϟ6k jgD&Au"k6z[NP_C1jcAJ돖gcZ;Hur\͍}ݼ!C;)Vjd@ nuQL&4NF)Y5f?\vF T2gUc=kZa0ˎCԎCm*035<lN1> ӲyXe*ifг 6.`AHa; =Z @*d46dB Hsmأȣ u$ -<4IUq$Nbamg @RW|\OW N`atA>D&jFmCVj5lTK$/c&ZaCk!ݷ3B |6!d _()-!>ـNGms;=v O0וɤιsF5cTI퓩";,wLI%s- qiϷ!%al55_~?ޡ=lP^zM Z+C =kZ ,nNG!^[I&'dΤ1bldb{+nK7LY-jtxtϧ_9&³~=V>E߳4:N}$,_n'iV tizMARY-ň4^Kg<g{@4rg >C! aJ'ԳNqАd| JnYC M(Aks}J>UAvF|cd 2Ȩ}VٳZWRUhfa0L} LwN檮fcMK:Ewt>qe'~ƛQ3_/_Wd"NYQdt@_W~ |zETN67x~N0/܈c, [@,i/HsB$кA~gwZ1b$h !v[x{f9o0Bj<V"̝ /l}_k`xiF@'8{|Йw:N9dƘ1A2Η9an%Evg⅖}o|sW;< Cj=n#9Rm0؞f{.3A&3%[/Kr_LɶRR )F7,p`9{㭼~`ގ5M &J^/&-~kl0[☑{0\UTF>X7yav@9,ܽ`y#b3"U^:[z%o`fha2N%7l4׀&bb2jd )}ӷߖ.Kk/pwWU8gjd>.MD)UQXɪT123L LU]_ZFW֑)H0mYh!aFr % bô>lls]2%a1*ܑw$ql)N3/B d[ x- 8._ℓnF:qQmB)ŨZLY@ 4^O˳YDÕNb2猌5dl~eR -`PY@d1nZztq9h0Uza]f ]9^nҸWXAaxVƂ Eq*8s69=*oѬf%RH~@9kqͰ~#1y&JFxwp/UpEVX Gċy"yiO޻u:RƛPg٨\_Ԛt{H!2,2/[3Xay HOAZc/cmJ G)6WPfCb1-n4*-4z"B T@d)\IaYi^K:m- '$[[ T.^No%|Q1mIpobxrd ;H*b1~ I-a..m`ȇ[6;P; 03=A 1崀|q0!8ۀfJ=G_&thoɟ_֥{Q[U?ZH'˜ ߷k=~rPfEyc_-o7U OR7Ȉd3e6ˈ.EH87|dZCgzo]U[a5Gnt['F\bHsbe C&d<A>۴jt6:dQsۿ){{ = 9y$a./IJG.R͘ /ЧwlJQOKJe(uT4duRzoP&C7D( nG&A P1ÈN g^veBMcv!H|ބ*B)k:a ɹI#s\.I^LsDFF]E?JHLjetcR T"ޞ(4ϘsR:5gvW=XAԲLiЦoPJ:5Ow2[1;%::s[.)i.BMTIՐPPZ]}{r^/{vb ݫNIl+>oR{|-o!I>ouޮaޯKר $du?3ktpS=B˕K $[ֿ9p 魮)+@<vc$J;B,{\y|EH.%pH)(QViȔC6߭cWJ=.?rLyj{][ß%ny/ [煖_s Rh}bQL ܾߨr/*q؇!ơW#E!廙})J[ 2ٲ(IW}ׄOXg WEa&jR}oǶ'(EV%q>ˉ RzїUܐVV0-]wFJ9c+RmCWoTGm\47fTzs􂌎Ўbf♞vaa2rgfnNE6-j;%9ڱu^Y%vMzWާ}E>1ڱF-ԏWXEfi4oy' N?v2yv}WY7E^e`/}^dT[I2L4S?*;S>;hxh+1J`qXu7^zY{\/J>}m\y6:lrt>g^HnH.n{: U. xX:hE-Vק$xqxkQ&<rӌ }[ץT6oT/=ݣI jsT$xgu[.?V0 :ӧ/YA5 w]nuȋ:o q]Fe㣗uG8 9,VT)&w-P@ xZ끈a4 er2j\xgI)Xv<˳1n SFӋN<`ɞ̫__,g Wv6:`w괧3B5ϷC뮣IƯ|ŀԑ<]ӫ=@%JqmTF.kqMzQk͢qSoӻ5)Ngrq5o06̦E>[opWKi[\Ͷ˴e6-Ёh?~kbe dGz'KhDs(֨CàE#va@!'@nGU~qrVqJji"j.VDtȷs-4ѵԴ-?B&{:EErIadt؂ꁫMMdt,ǟB)].%@w*B-J8VZp+W˼הZe]NUhu|NR8ωPJXOC,n+L_pq- uCZo({+Q0FD@_8nl{7X$(}(O1H.h+ӑڡdhJ&N0l8Vi2,Ԭ~ŜMwceCj\Vn(؈'cWfyëYxO(8r.uINFZ |a~+ః&THDщF[x58vbiY@1PO^v(Tyk(/赤WP-vXKfb%ᎃ;=$?+ĥ c*.D>Ց *A}dBI ~ƣ(*%@U )$tx!nMc {/WtMrZ3rƗ  1 0}AMfzgE4 @;癕^/qZ` V}د9>k$W:k,wE|G.-9jPgiHRȼGɘl ^ U:hSPp={ =Ruߨ3J>  Pj|z3 bHKqz\l Y1-72Y^:.GѤ~/{Jp>[C(~xF*,+fYAʒ<Ңղ1C{f6}ME=ty{}{΂6}#tC*&zأ0+deD#m@BrJKSy3X(r1!gzn|6ѐNYkQLj,gtgĨ :L9\/y 16'FءAݾbԾ(5wK~{om n:?O(-3#*R{g$3"FY%ˢcI3dPX=695ɯSa(}3WB]|LLUϨ> ]a X2=9 >}7̴StB7t$#s7lѽ}! 9Y3# r#sjE|!(Jx=k*0Z-R!/PP^Ս:*L"ҒݽxhXYo2VU;`[Q ȓ1|$Vv]U%NA_jVQޣQafz]{z+? &4mMŰSdnٕgN&q P5@!С DOVrd䴈j1\AK%[ʶ2,B48Vctj6/gq(y>ZNJF|d lL۫[!._C% fHS 2g&B:.ikRX,α%lT~#O,(BQ2:V@6 HUEժvlQ Frw#N&\!|g6fLD(n=<[&-݊Hx[ugcV# JjrnQ;GU[( vb﭂":ڟM`{KT+FYŃK^ z]V}..\M//V@I״$wJbn1hԔ$Jrn PJ1NJ9F,~8g!cE~)bh!(c_zn 8w_ί.ZXXLb@%fL $)c_5R(#Q юa!1<- ejvs*׈[>(X{-%W<5wwg}̎ T}2*>6{kQZq˿t`s?w28`hZᎊOf-2ĝ l.۬ҳ^Gp5ǃw 6\4 'ؘ8LZ*R  P!1-8B1j4x˴Tj)n`EZ2٥l㸨 J*|2P~5ɬ%2:6" qϳ kF*V* Ra4euW%uj}R\]QtQڪd *wBƽR@]8gpbGvGyQ0D}߻QbJ/|O85L=ݚ_yH~df#[N°x4T%̂0%AJFfEOf峟Gخt;, c$bXl8_fIq\yb9_oyU|HyING_e&E:lܥ ~fVXh\UW >7돬|256J.U0 ٫ &wc/!`شJ< ʚvspPY}OQ:>el}IAF}χK |J}kP9Ň9$͒bTP0CJilTh3'F5Fk\م AM+}I/ze!/4h K aGjܦb޿M '9985c6\e 2=?Z< aNNqC,D3}vJc.\$^ڰ'מ@/z*au,͉ DV}すjY0[\F j.j1sҲHYo3[#h쾦3 VBl|O{BV|iꑺ/as[x$&4?a]6s1f*~ns2!9 xǨ/ FVΩk za\8HNa~޼3CڪoN7% B#͂fbTP/_ws5n'sm<$ xy4U9gE0o[_i$] bpV?z] R\jdj?ԨurA?>/5*`0O~ ]z ,cA=\f 0j==Ks7`*cJPry G[nݏBis/ ?A)e,3^+uP(nVU;.> ݢ{lyZDN#%7,r8(%?[  IALx6E+eM?lJ9:2ci \Ps; \/'c> ~7;O9ڼ";͑GN;!eBH,;Z&ٶEv4Ɋ( yȕl=u0ۣj4gR ,'2&M²&"J.pCS L1Ny4Ҵy=Y97qB&^Enx`(2?Pz*@0vj'1$pZ+RMx*Lcn aPm|Ƌhl~geRLs_FԗEJVT"Nxcℌ\Sf 7-0O9% y,3۷%#cIcLۈV/(0pӝzHUc?亘a_Ck &[*=Fm.2AQQoݣ $CH:@*  p$u5E*P-5fɃ~@H[ Ӕk4b !Abo1-859T*i#uX-cxZՇ͞Ld՗;NcJ.*P hBh^ '11 %a'+UϗS#S o٥9QQ0UJXX&}Y-*@ buN @ $I3U|OF!٬H8VpBӈ9 J&fGD8y]c$2>,w/SHHETl}7*S\R[7Ѩ0:}rz+bmY}lP _Wk YBWɾo寧b_n !-I4SxOLWLjGUѫFy7uUr.Ϙ+:ѻS7'9}rKYS=%Җjbo$E(s@ f3 5ԒJǠBFioȤ9\0 RGOlaYͣgPƾ$[M@sy@X+(SD<>QHd|#2z,QRje9A,oı*ڈ"! 0ƎL 0F0%PFHjH}|#k#}'4E[ՈX :oXqpɯGQ2 N)*whB񦋃h zz ܳX&v$B#as$T 1ySIΪno*~W&-r_r~4[y[FlloXEZ*dt-b4+9lV |aC՝+)Ho4VȸqQ _A4Rt X^11b$fcG ^#l_8,$?'d]uF|}޵79Wj ћu1"ړSt"\cW^\}Jх>ciLsڼg;zFc4j}ͪ}vy DB_ &Rb}_c}ST@E/ s' Z5ڀ` eҚժ%dMk}%((o š:̕o__I<3Hlyh,"2W&`\%>^%ܷ6cgf ylz1"ӈiw7eR[zO.P7esĥm1{Pmv‰\%w0-ɘ=h]H`_FI@Q!^?b%lP[YQr/{[4WK&QP %0o@vd rrŸ BB,.{یlΦy#N76 )ɐWyuC2!ou̻?-H="$#T6gV*!hqK )uE9rГY:c$:EsɤT"$TEB1tRPY4>`ng/rc1d-XNbje|| b5~yV9~sүFyŌK}唒//hK& ]2Ha&<֟÷ٻ6,W~ ~Xdda5R-lJ)Z$[f_H7;$U_];ĬInŠvBsI];sѼ@RD2I WhǼ!wf/v6? EG%PQbX19haJ5L*QzW[Rۛ]+XQrCgaR)p79_|ѳe&\iJ8DT.)ECaŮNBF48 - +W_Rm>Gp\_ #𾿸xrud8G1q۰ Wv_a0O|I\ )Lj`lC_(MCҬϘncg9/fK6dn2iGE x9s\o? Yh͞S0A.&9>[6h.i=DK}'.}^VDbV)s0`9l!fI I*isBIbQ;* -OA`\;mD&h".bJ] ?Zsί0؈X!eGNBPrݙT~+χɬ@FXJ4uFuP"VCĹ=xg7K{LіlIüIetrǥq_4l0O.d9әSwdDFMc",T,0,OIOKD4ӐzLXd˦l菈K`/`L5¬坸G׋Ƽ@w@h^řV #pZ}ê\ hK`aDwsg2ν??2 lX;e+xE5$j1/S JMyӛ٦mJJxdٴTIq5VTy$n NuH+Jk0V c{MVA0RЈ"" xּX#}+a^o; nƭ|V WIa ÐjWfPb`A)W!4b[.EBk'\gqWW!_3<6[t%62`U!x1 JH,CKV-%hR&rEL}$m\ȬBZ_i'" CdRbPq΃A77:E4p-ƒ/ww㖐R*amI%(]$a51H2I5df0f1\#'y> l`{dr_+F dXS$ * c8 z=3B{$B$i CwBy$toʮ;gC>M11"Нy"8q!JB fy8zc7J"9Aye<(E4˒PQ+BF;o;#vbXh%4l?JDEw3'Saݕu E`cx4a}LCwx$9ݥ=A&\ҔDR%"t Ҙ.OZӷQiJM"~6'NH=<eEdeÜQ~Ֆ'p_ R)q$ tDRȽ(J┧h1{z}MO 猟Ư;GK˚-+U|;Q;f\s|Bc|^vn[ܜӡ8tRBP&QM#gJ%H[AV33k(jb$Js(ρ#r/{\Ȯ8~#z/McW Od)A< 52!^!,T>m`i>fsjK A'xK}E]6ct$cKđ& AL W_ZP5\yRX)=oK4crbt JsBU./%T$UBWJ ]$MNlf-#y"ł{y\!㍯"-F`6NR)%5HF"ic0lGb 6{S T~ݽʟFtbAP^A7>_BGv ~qc^:Jab "*fLDzsr!ΈspF(Kރ3|YMՀj3!:ηźj.[ӤeRHt<5Hax"S ",9O#r.w]Fl5>d=XdN9Nje~}D56 4k1;4 .MDG }h߲_=Rr؎u?%%űHĺ8&ŹE0ʌR2W0CyV-" :y,|`{#>Ą` =6}Ad!:'!RҚcv2:lf_%nZ]#ĜIԜ^ W'}is}Av/qNzh\ZˠSU$"m'r 5-F`/yVβ)^nRe5DH^M2D) hoI,q!E¤^ CzP'0;%+c&O%>06CxrcGrwpNZͳ? Jਉ;qt'ڴ@V4a5֛Q;Bœǫs˵DhzR+?z|I&Nض gK*rҞ՜[ 4| X|J,ݎ#6ShֱQ3SVJKH dt%Tu[VQA9WE:)X(d%!R(F %ҳp152cԢ&;W[qB7-J0nz%AȸlVARMP S&~m:$@wv1VY:i}t􊭁OF랖 ~KݗZJg1`^i& \]'}4^ܺ./lVs7A)rc CS0I ĽG\Nx0! qozcp.\z :{ 5Ѹʇ3X1x0lL-*̀SCDwN*SQ"I\2Rn0#*bUh^Fiṧ%YCp5BECD5 r7u4(9?=MԆsg_6g7S{vë͆-e2dX(Rb׋EP4$"Jy&uчe<>#zjHTq|*ddK(=!BsC4=qWŋqTA/ţ@:9'Ko!`ŰP%ռ\o8͓UHJP1p(}48an}w_DDa9Dʀ am1zvO1/ҧ;8EbUJ=XS"9ՔS4\q eG;XKvf)`gS$ T}5P9Ĵnsѵ$EsOWT Jza'kg)mOVLSnOUS9I|[2B\Ew!DB5&0< 8".$ߊX>AUb+!'is_=1\=o(CϤV͠ F==ݞ5o,|]c*/W>C:7?m`=lk6p˭\_x-٥4u-/#[. Njp ?pYnͿ67tVG7+1M74?n2=znCu[lhȧi\";'@e$͋Mo;RHtEb6-x3-Q1mcG]"ŶR U$>x[x W񮬁ޫ;ѓv:G[nn]}7ƥzy, z=t'{3q/E銃OBDBc[1ْH\~n}$i,]H-nJTD}=tR+q EubV"J%\ >J* X1dĒ&IX'|T0mm钑Td,P5|$ d\"<\5^gvt27QL=xL{7a 86tDpQ7Cs$犎]*n,'d\k7Dvub`!.to&A8{naQ٩Bs;`U';=IykC@U< $ $ԣgVKӭι-GI~+x5/Szc0=@ tq|]oR6]xc'ϔɆmL>H$vl4}(S ? L.L{gJD"^4psmfq$M1>lf^A3;j<& V|zY@y9EDd|[ieOl4\Cŵ\9R>iaK?M:RJCSNGmiwBӴ\Vr_@KIMejQ9jykr|uz11ZU5M9Z$ay (eDV(v$3ۿU5U:EK?C!Y"ĖШ*Th" D&LuOsjd5]56q&vZXKS 1UoI.rm6u1X DAٔLaNe im]zxn\1tjl}%s5"xzy/˕cUXDT&SX6~w߁:1X㉵`)oћOpRE_ś6Dc4 2)d)d 2%T'K!w\R T Z(2hQBcVϴ=;Cc>۱~^X7hyidVՖ_I!;(Kn)-,*#"'gEu-)`FQ{,FVіoÀR=D ](Uв[,5`5ሖ9z:&u 1r|R|,%Im kUЅE;P(sT*<_Df= R@*VU #,i<$7h%$}0;b)cˆ/X_ I>4"8EST)iURF. _R!~,EzΔ1JGIpgGqu"zc.J&#l(NxtLE!h>:A%0`qDt.4! 61~ SLϢ 0*}y?Fv˘/{ȻNܱS؇U1ĉ dst>B.B0* _XTJֆi-as763MKAފlc±.Knؖu^uY{(9u`+#\ XGC.rvtGc91C40aDK4bHD:Բ9[zcjxVDs8XBl]I =9ƂS|}=dܖ*F*bnuv&Fo_nJo,:A}8?|VA>3(txUF 4|+f7s\qs@4#{VSe52>@j7@k-@n.^ h+dmR.Ozk9M1-^$|(8d.^cg`Jl1@]?~|'Yig YS;[=b%c[UA45d^LW:x4/-򟯯^}{e{Cz2Y~srNbT.Ży?~cΩ O/Y~ 9EqQܦP5)sd,A]v l!g$>+(fYKe>NI|8=Q dYJ94F4m@)<;:Z~,*i-6XֈuޔPmaxBl#KZEW"U.hMhjLiD:FS](4"o'.Q4+UI|, w^12F`L4Ȕt)-֛AD%򞤚4tMбXaI'B&PAXƢ]X+Zs|*Pp3*I9źLzVChȗRl~ \qdgX!Fmv֓Eh=9WfbJ&R*(m d7,ʡKm΍:׏hH$'f1n9l3׆fOsNwHa6ȕ@X #tIa#֠EHb0ӶaYQfUN[ _pt?..}89Q-@YIZ]Kw;]MAhV2@|Jw\}pkr46qTOFB&T'^ewVJ)) .XJP|ͱ5Ԙ¨6m!F7L|t9oq5[oN!ɊM1'@w"5˪; oeQГ6ڱVEXcTy0^.bdE6h*usIQ Fns(y* yTyi.sЕƐ#Ӷ*UTri9=(μu:*_?{P'[vlP9*[*6%H5t!F_.SI*pA`PͮA>r7#bC&\./l9n /ZkͧιܠxpE7ɨ6Y8a*^ .]|dGeI agJ=ݵ3,4B-^KE)G>qVUrYW렯Z닗25G9MU# *ynbҕB9n9oѴ=؊/[۽fVjB'e_5HӴtM^*<\mVZzYnj^2AY%>ߴ •"0-w  A X\Ll'wNNϚ)2ڹ+ONw;xը̈Uc8D6}j̧]L. =r¦-Λ%O:iĈuuh@zW:sNp%B_}_XHԕ'__Q[{(H~OWX>R/k :-ׁຑw\?c1H[o\|:oÌç_$PQ"V5)нs-09?VwcAQR=#IXR/g["ct]^N?QA[TpzVsT3~ڲd9F|x߃dFGD'l?,zqnI\Bw]jnA! _GnǑֆHÏyvFcUt@:bd ]^r܎t(wdR "d>L-C9p'zt Yؐ!v\Q1">5U=8` Ι64+gH㬎mas0ڠUJy:lF߽bmC K Ѓ/%ZsyO)W܆f?~u ?ꨭP2Uwt)p: Jb8]Sv`9Z\wkξPIN@Ɋ6,~1Ycn`y:S wǠaHBD q`>#/NiIbg4`q`mv΋D7.Nip=,W*hӡ@ev՜_ݢ|Vr$@-osa._ũ$V2goO-ϼ[_LvbY?3\DdNks :o÷DH'L҈)po.f -\X^yM9s%-I-p5dU{ѬxkyCLu3 eF!*.Yϟ~?O}ನ#vLwq/r f{q cZ9mZ:#ojjG9O_k;]#~@څS9Mm\r?D0q ^nKB(B.Tq;5" zG-[w>ЬS?zq\\7?Z[\7S_`~1!.uMsөFOѬ⛣$NUqt{ T&RjNEސ-@AS1׷< M;<%VaΜJABoQR=ؘ;&HQУˈQ(>;FuJiXkC]X\AT'0_]j8&B"KSեKgL㮀ͳ3Qީ Ov+E Ehʛ2cסH/>sh'-5{oϣ"Q<Q{yZ>8zX"7w $3t(~A8 שTn0,~/My\noѳmO>Q,_}bT|~7(hJR*X#ъYDTlL+aԩS$h%nBUB-\t%tK)UqGv ||*a<76s^{WW[}O .:ѪiEa{@>܇.u`NN_]h!Ah+E7-I@,@U3hŌD+fhe875  T^~D )3ZpLSt`+PSהpyF )d{=J(&]u]=)*Q9WKcQxA9Me7[ _݈Wnr A*͵S3Ti 缝1>o74(Mq8JS(Ɩ%x.Ql9۠*˨\/t Rs 5/x+bq]~㸬A?pmT'.zqrɌ;PGMS2|ʥ Fp7/ @K*>RRsQw@R+.Z J;s>_U+;K1ՙqoa]gE[VUPű pONJZΒo@]yu/㊊JWг%UE6 P_ԍ SIG;p@W~_ Ҿܠpqs[t:QI gkRިjJ3-8*VTI#WoaRC.>{ף>{;yc ɪ}[.7>@t~"-t^z1tT^S #Tb>Oxy 2hjd7w>&%+⠝BɩHinԹy\TŪ81 MyDVq)F۬Ku%d_͍aL8fF13pL")qt<6:0TĕN~wrC!P董XlU QfMeT,\9CjVJ1qYSjʹ4+)[t:r(僔g_.r'Jcq/`9Q4kGPl%H2(=h>%~}Y+ tg()y`][x0~?n2E)xٺ]]?ː3[ܶHu)m +lk#NwKL̸ 7%;(g[:6.2 %W]VUϩ-\$&TW#&6tg!QUrCdtFy~!nܘj( ;:8j2:T!+Cqg59’\Q)͒a~$>g4?{G¯e2SJ)E?+.aSg׋ӯ_|8=:<歳ּ>R]j?ÿ~}8~<ȧV kg}=+2|iCR7a?xW /omewO6 ؽ_ތ_?Z͓-ΌPk`{VRmQ<"be7[[fm@GaQrsj& Uw:(&ATR̮syØSUiH.pMm)v3zG^T/q\ ן>`ܢ#+0juBx%7q-- Uqהw:c```ɻX.P}ݓ`;% f9?{CMtY .Kl!PP͇IɡbdYc6, \VT4%~K"T` Ob3\*+XN:;'1$(8 z@ƽnv诎厺 C^LX'?V. ; ҫ Q!pٟ^xUaa,Q/rHwzL^Ieλ5ϪiJ@ct!5ŢTLJ.1lxp;{ђAե\-Yx+3WONCW8=[t9 ʹ< )yMlulR,*IŁO0bMALMtm SUkK#Yj !RgAdUgǭ׭z^şG5:>|bξ\\-jT``Á##o꺹>1o1[/g*·h?yh|%뷋p==9÷`6u:nz!Xb^K7La~_-5(Kw۷ϻg7X ^S(Mt9 jR'P TR ,@BbqU`щ> &8x|ru8· MSyf JgaM9n0m`;3e%VT'[+ V\`}k> {2|͔3 +kLD>nmJ $uZ(TɗT+\Xb|etG ٭亱p'WN?[?qt?q=oDs r f(*(IW{PnoL:W4u7d,Id@rJ{k͹JnϹbrn y4R"op3cq7X*u"+ o|bo=K`tޟz4{7Ŋz$x4^κjˑ^+yTcǑg) .,?kF5~c9) Urޮ8'O66k'茿^y9VYeGLcO ޅ}ț,q8%/Q39*Sz9*޴8| A+ZNʖ&9ƣ <}onwxC^wtBߒ#M o>Sԕ5,3U񖁞mڭbBY\Ow^ y1A޻"XxLtpo`ፖ8Qr",ODp"[-k:ƣHOֈ]Ww"2]`Q*Lȱ樷Rn 7՜N,/e^{~jS;]F5Q;u0tiL!uBjws*OL6_~'{g ֣]^>[ʚGQ`V^BP:ߘOh]rh>:x^i=b֣n.ɻͺw-em?zsU 2;}x>a7Sj.OLu d6`;sj?_e(|:M{c`N [7_;x 7+J<ޒY-N-=Ho]ܣ*0sb[hLyK ⟝ye2fq/:A@B풙0;Tc9dYt5HK LE`Do7R eBkrշb2^'4lSHX:'j됂}SPIzxst83^V]cuq7w}8PJ%7l÷w{TT啴)6Ң]WxRe8IMM=Fr s !Yj N6'Ť i'}iY <d Cνr7W)f982q* b $>*֦Oj9LbC,>F73=^Tëڇ/_r'$D8\~l {ih!۲ÉRB̮&9\r**@!r9)FDCnuMj ՂлʦzKB ɧ`@k$WUQrC%PѼ"qXuR$IIUluH$Z5ZFH߁MAV jA/jRE ~8cGC[]iͻH(5=ּ>Roב_ߨp?\ӃAk;g}=!#s`[vrx(OnS./$DQ} %X]~1#':zW185Zm];l**|ԝ]mSSsŦ4+-}rQª+z Tlu¡v%(dm:t2#roL#mrb,0럢B}pVlQr9{,\P>=FY0d|U5\~Z)\)mdUBQaz?%\">b3ZnnFRDGl8v{g.(e[lq $! )YaGmMA\?(l[ U~\bɖ-&oAj7C(C'9|E\!%'-*D0BMK(Y [7[*DJEXݸTmUmekw,Ou9 _YzCz뿫pC *ޗ[EN;T彊mˠnK]&dJ ("껏Mt1s!)檻ba7S.^0.&E:(:T]s ؜ɪ|kUdDfakCOzDH똃«eq=07Y"?cD-a?G`|{&G..AF&=`O8!Ǎ)p(nT-N" FVʕK%f\YUZVZ$R bԤ2c"|[ ?{f%=jTqqqENPGWR#Z)i13z;t]஺!/̘.i.Re[8|&(f-%zesY,1\V3V5;- Kc%.LZxtamCNo8H#NO>Z\٤6c_\P28}2D[E T:M< DJZ$FEiq%Y3yLw۩ U*r+>7"ULH1J:9rq eyΑ_PKF ļ_s #(*VyuiQ2ZXHKiLpBs[\@"EvC4`0FC=AD&1F(qbL *>Y1@h6d_%+QBsr\:Dlkۍ#9"qE-"2؏ oO8^C}5AC-j]+ޱIoM$ #Jyzpr.յEK8mz/kl\/t-d8Ęr5_@BQ岅)L apQucRuU3]ЕG-/06d،bK!s@F+}ex^Uk]pA9E#TO6 XӹgTTB Q W峕t5Hky"0FVtvi<ڜ{eGJzɿ1__LK/ 9"Lnv>'+hM}80AЉ 6k[T8]"1b-^S1Z JGFvFߘ 0}y˦PIǩ5nWzEGWZAEx l䓆8WUL" Y{]q(U^(SK'(?*]*6O1*EUiWLU.R`ɱe25ʺ|+И>'8؜: m_}Y3(JO u`lX<(ҭ)*t@& XqM5`>õ <v改֬n&KXăTk)WJoܐz[3$KF&&",#urQҥs 3SR  `sRRE;@AH8r(e`Md=jjLߟ0į?nҪ(~O瘥+ /ّ)JV3|'pu\+\d_^dz]:y []N. Lw("$q2%xIO\2~ȐIr&n5fS# /aĩr'k0J|,ѦIjQW[ [T-˵ԋ1[(F"b4xĨn`U>'[d]:6)3nT-(>U%m|>_㉉Z$>Dd KbKٍ"pdUsXBn*֍5-hv q%*Ʈg&nmo\K=ūv#~ѓ'ϵL/Z+X-J+*-&@~ASo899`Cj{ú6 Jѱr {\^KԆUbHK՜s1ah*}ə{t*)AI\a\6J0nUi /*y)f0ٞ2u\ QT!DS FxPez:m#HdUqpJ.ͰyINBs $l'UFJӀ^LʂV741b\"IFMHaܷ~g{].'זeʮ~80)g#oUһe~:-\@0hMm5P6n]a .sto [uV5vH)";Q7>~Pw[ͷ,ԫPoٲp#bi>/a.N{̠ѱkI| Y.:SNRKG9F=R tcCր㴊$9b)X@Ī1 +@շ1Qxbmd{\G)`<ܼiz:z3~ԩ2HկKy9\vr}b/XSW|j6˫U,,oKZD\698azʓn-~.+Ԟ~`G?s7B㕞 >rУA[?e[{(_X[W㎝2mPAv[>C^g Ů =WGϜՑu=Q ·cn$ϯn*):詐#lh~qBa۩hFԻ>(9ZЋ֨q%&#>S`DXr[.4GLD)Eokt{}=9"UZmCQ.3&t w,/vB[,|Q]{F0oE[Dh &Ǚdeam. ?xo5Z^X`X?n%oN*9+睷yW#is0}3P3(b/;/b^zRI3x+cYÍ$g>ӑ`Ar[1zh3oˣjzpS&xQ~(,R;Q HrFȩZ/2Q=X(e gN\Uy=$bLglXHތ(rmD9n#~LKdg@D$2u9̛Wb폺:`.[>5k~tghyye8}Hr3C/$c MqN|9nwtY{@ šoaE=S y ~F7bL F/i =>0b 0]K 1Fň|a.Ącl4"\IYF\(&-]Ԋr WZ D{/aۮA*7}z ٷ?r&`S\a^?߾ۜLղJ}v},Q1PdٔRZ=h%[ ԇlVu劕n+X#5D|Lie;N&D|>9M8)R 8DL1Yu#)UdAZI]1 _KUvy}RTb @UV.XMQT6jyxP H81|Cmɲ4 ~/%j5Yy 5> ǁʖʱ֞Zz,ŋ8rƤld-e*6LD:;ϼ*1kq\"TlpcǞAG 9>8FZqjŏOΡk!@o*=>k.i+_zm^W=jko}ڑE/UΉV` f%NR4D5KRn~s;O2W6)g6†GŅ7}FYqD"B| vpm!=vX6sRbu6s=(l 7(x(^J9hv6}Xb@;lm9 m Z1)n@YC;leŊY2qc!=7ݣA;l0sRq7͜}μc#gs@ ti@ga%A;gX-<0 vr3(U^+B;9Є2m3Gnδ6NK#_;ƾ#iB"u! GvyVQ53|^h>Q5q F^rc 5YZ)^GR ydc՛G X3Y9DGmn5WdPB(t5CA)ՅcFB#1P FA1n[6>xQ*l;\FswUj4:Hvwi6%˚MɢfSh6u'Y0<4tl#fr&X1XZ v D=dàcjt 4c,3,Lo,4Hn]fλV6u>\az~Pnj>/1;|㺝6}O,xY  r #|ߗt[(UG?a_O+ڽv/bp ;& axvKOrs?_>Gx ]oGW~ٙݴX!@ 36l'vuHi!)_A}MM4ڰIa]6>0˄ -Ha> )|P[M$Y~V{"\ Z.k&7ȯ/&?ٺPK#Q:;2Líht'x=յJrQ@"Bg_[+[̑4E6WoK< g-q7/Fd|FjWnڝfuԶoe&Dt0ތ]8֘5Yۨ1%{ 0VX+}xdo'hO4Vxhgn .I$.M(\i%@kܒl.2AcfIO#y51:SvCcDb&oѤc,&MZ; )^fj?[+>Ob11{Jp*D≕ TQWB_)uոW7)9l N Y,Mt0! N)꾢 ҫLլ"{F!$߻卛RmMȪtV)Vc %P#uA[QHjWv*ySWqE ??ck!`0ʎ&5L")&ZWL\oͰR ȼHMĭ8ɂT:q9@FK$$ -ϠY) iy%,1ט,oZho &~Bq_L5 nh/հG/s]D/XӒiqūWDv7PE{jl0ӷPeApZ{n fFnlYTsy`0o ߵpoP<-%VXJ0<S-QLT)ў!J%ㄘ@RF~2JB0Kƕb.y0N&@ND`x$x^hQB%E dX) ʡ-J "+6}-"Rʯ?Ҙ1alEy 4|df=BBr5 67*Ϸ\\e:@$xv0o~|9kltsJxN?=.>1!<޴N ½jVƓn-N-wzf74ȷϷo_MT,Oہ!j BK~K#]r? ^fnƃ8.EX7SfQr.h ޕޛK{n:|Q oFM=o٦,{ noPQ 0 ,6K*,X3ܶwvt@iO]{ЎK|IUPL#鱯*?^*H6$uaƫ;L\p;1X0 '&tvX?+L7\:|Ľ2s&s\{]78nq-hq-kv$BrS)S%fX(z@HNO-s re\92Kꛌ뼘Yšn۪Qnb@`͹zq_ϬL̼lڝ?^dY})۝slXtp*SF8*%9rJXTh%q.Z1hJj蒃=%iA"K.?,׎$!E&`qyBբ&jzfߧ2( a՞\΁0 krI])@4,^<\@ =e F"Fi)% 8 ևOcʤ#Z9;TEriq=zx-. )9K.Sj2V:NR CE3 {|,4 HEF 2`gA]7}vPkzNgYxj]arKiAS u G:2'28pĂՂ$̝>$ELH5Jq8)h$qV;B;_[~96ԬF[/ωz]3X&:_VŨu$JkS&I 9#UDFX*F$:0`/p \,Uρ@6^xnJ`* )iV#x}y9O4ܞº!Fe.$xFIaMr1Ѓtb9g&U;a#@%H'xJnզf6"곤PǂXl4O`1 ZYǀ? v?U"Nc;/3)?HhMU׋ZJ|@ j"ҮPag/ۻW&PLZB#%_f섳_":TF.W4 "ղ_+e$ܞ(jlo{)B29wWkm> |)0#Fa3\1;a喙;AWt I6]'ƽ3Ȍ~Zbj:;A_|GzͿ7$w$t{R7i Z.'50WLyHc:Ht`%];Jx޷x)hЃv K,^$8dlv-a4ܑ] Ư50GBQϞ2j.?H)eĞJ^ߑR"%v85r)X'aˌk!rC4[=g5TR#Gө/T+䐕Ρh5 ClwګX rO9l{b]sv2!1NgHS*)xU9u*^J[Gxw }A; 4k,3@1}{x.Ӛ>,Qt&]<էVRD |V fT(].}3͉8/z`4~s>=JgoVsSδ:Z=G?`/L0Sܥz1R+ase0C!X3}bef] ?םFLvQ} oz>@as}%ޭ>a~qZd,k܄<7X")0EB7yKni\&&nkVI9wz.i .t;]k"Md IEwcLCK*mLn7N, LXCW3LR r]+Pg̿=Ĕ+Tշsf_O~/(!!g &*e$A\Z(E,O`*:4Wx X!yuk \Kbyе~W# s8.9l=(`'1˵psW)q1@r$%b$Xx%` vD#y*(3oRGq~X*/Ӛ>,48ՆԢ Ͳ>k.OMB"IyN%Cb_@ӇH~K,΂"1(r"1(r[ K,:S8ǑBFyK\Zi)! U* K©.]bnGJ,|5JjFcѽ,yZ5UMr\faCs,|Me+2 hGl'`6STyG)嬦ifS? R$AƄβX-CX)QVu@s~%k զ_5h ZؓH%yʔ@աbH&vZďvSmLJoF?1Hj-ɻ=o?&sxfw^׋5\JԘ ^}םz~ZND5D;n.X(ljY``k$6z:`s.ZBOVL-|Y+sg+BA2w lc8mhli2㼁} czrC/PE$<%ETɜɽLT4w\Y#ʢ!` гI.qO|ҁ?3c$XmOlfQ ik0Rt#:Јjr R]jgz8_{hy ϱA"3Hl9!!ťCJ5qpښh%m%:]ܫ\pn"$c*(3/52PSxJq_`X)//rJWN=\O_AK"KkWq~2U8UzV=Jq8n&%OuVz`ݓ:%۳Ɖ\ 4'W`FbD`'z;#ةz$xjNrME{I VcRbiܫ(1L 9NhFM_5+@m:|vۤfQ:|->*6v۲I.yJB.9|-czG C-2^`MPˉ99 ,6I]K^ -?!$67K#5I!W汨+ e!B\#Q@Rq8Ʉ8% !Ĉ܂}E^Q\4"I'7a(Ea~(q`XR ?@y!`a̔Y&AiK0,ؾx\ 0 $0cSaH( O;\7nc~VcNiv+, :xv)&1 \iZԄ*vy*!h⃵$vn<Rq `I(!3Pb9ҞDX4Eܞ`ϥ]T)b, zey[HR%x}"1K/k]j)eٝϜi)n xS!c) w1((3y͓^9S!shccxg|CoKT 7E76oO~>ӣ׽n%>>9*SF43tTSwLjx b`׌,9h _қ&@(ޙh]X74c?O#v0aYD'8Q'`/Dq`#&Zx-%т[X`% I`9 VXjX|խ\u1wTMqG!݌Ҹ$aM4Fqt> 8V؁AkjX#B3,Ƚ TR |8g@9W9kR ;.:37?VG3ebk29]9`+5f;:((i5*  jdV9e4l< !HYu꿯’o e%X&He(Lluzx̫?u݁s%"HKB69w`5C Ll38+)nvKn)ƛI1H![)v ,Uܵ۝5c}8V*Y1vxթ!3 _?Y%Yn%4ct+w#)}_rH Ճ8w/w%&xYo[Nc,nF @*x0S.%t%K1냚thKDKꓤS!vWkpgzSqt)r?ǦrD s(6F$Ԏ& FMzb8T9g3 ?c˩ %F)Z]hW$<}iӺ=`X/uj)&SN}#JXе]l}5 v:OF];άMƉ s><('3b0&)NHgq0z3bX)Xht5>[UĖ5(PzuR]vFq&+cValkآG*;uJ0Қk( m֐jT5T0\N'bF@5m֐nt5T; zҮ!e <ƃςv)L{=[Y;;.w'?}G\3@^StbluMhbG]JNnK8FKƑW]AiUXWX!o X/0EB7!J/I( hkWϾ4c70nIYArCc0|x-@&Von qֺo[J[Lۇb ؉j}ҧ¹ZDUm9D2lM~Ӑ\':1$TjHWbby.6G^ +tȭ 0{9$Qȉ6H&P}Mb=3N܇*}!6ۏPٽN gb*chc_Sb }CL*φGaqz M(9E-4w5YEZ6Mx'ٵ֘ʋh$}{"4'9*2 GSn@0-(?hW:?BO㉟=x,P8f8e[=VV8-{8{Y wC׻[_`X5J12/AxNZ!ֲg."PؤSb!~s#4?5LtddvI3=ap sŐ{w㳈oO.fII)wdtpճ'\ϊ7W{{p4ywt~z0x/|r¿5#0糿~_uG>p?^v'߽}lv=ӽ/]/._ND"'$PwAEƽ>7 '^ť鷣\I)no!`W1XX7ύ Y=;xL]Ur՟&ˌSh x®-2η^δK7?7nYgﲈix }=g>;)0,\|(T?}~`1=y=˗oaG /+euj1.gu6<#Py?/3c_!g<o^blM4\ ?[quL0i;.K W FQڗ_'@) =0o7 DÏ |J!*}P$Ǐ׋/WBLOluϟh 1?GVWz)/3&4M}oeߜ}}.?v'Xx}=X`HƸyFppѽjkX Շ:l$lCQh\(_b}{3h;§"7^7ro=ԛe v}k׷v}-H103DŽɘ&˵YV{r" z v} %,ZvZ[`[[2_i"Ƃni&@ˮs OIv}-:,~L[Ȯk[Ww8c+z4չY0g ̻)jex v}%ԃخ_zYI{M  VrE вkL1$~ )\231G$ա2G-~ 6SzTӷi3՜:jc22=pRZ#'@LT#ȧx6|z I\aH'3ƃȴT"s<\_!H́֐\.ڵ~3,І%6qaTPQ>⼂5=X5PiRaw~Tͱy<-DtcڂZTM,[փsVDZ e8Pz`_iHp k3r1&u Pr'%[FP"ޝ M9?w{iA}ݲRT$t_Ď.Z^4eVid,_W[i8r'H^9v47rW#r+g] c:9cÆal8<9!Di-Q .e:GP ΅ p)s*Ep _l/Oܡ9,'xפsԃJ3|+fڦsMh9tMIUt.ǫ8B;I4Y3) eet0X;v4 d"T$[K$$ [K$[K%VX12,ecd`+$d%Yέ%.avvJ>lucaJϬ]'wSqW[/l> d.r匲Dj(33Υ"O;"!{EM_5RbLQJ^yt]':1|ש<44$L ńLX+2[Xn$ Kl26}Q#myTPXΣVy__ |Ii4n~~|=>9w_l 9+iSB hŔΦ)( c{]o$Wrt_\r6lo>%X靋{zdcbό#! ]&UU` u/뜙K)#--ge"hmz(6PPÔ[ %4w0pu!Y rC0^*L {JF~xR<#"yQ(?kRfA\%ALKĴ$ALKĴQŊYΖ/ԆsM,h+*Knc^DG:ENr1Dc SE2=z`4Oo^[ʀě*ّ ql?tAF7tqK <.)c\F - Tn"dO,@V!I_h,%$=7(ۻaA6&o߹pw_o?׿sD+zLgn IC%ߞ\\PJ+-\3& pD3BC2Q!DG*HقK"C1;%{(^956SĜ].J. [IM<4pw<8]U>N$kN(s6kt40BT{*vF"$qu.ڻ?Bʣ R̊0kPm,mGH ɳr *I1F 3Nfe3/e-dP{q@!$0ו6@Zpf@ 1-1v S^xuB[Uu8n ̒˃Fouu d:{R"}Pmg+z2z!{4"seT$Rr<ՖFt-,Yn/Dcf -ރkY( Jxi`9a, 2g]%2bMau~qQMZFGL@푇.Jї9_v/7$K'y_K`CHfh~ۙڔKIiZh4 qNn ;^@T+Dՙ'3ՙ..\i9ȩ%0k$T"YƺD'ٓIdLoK&`eh C hrO m%vF614:؅oߛ `^%'u.d#-߀|؛L"8)  2!XSZ e)GNn%UI!t4#l\J)DIj ~BjCZJg=#haAow۟)8H/NP@0[ߓ8{.s;o 3clvKF%07#Ҟ"fRjN76!#[!i".]YFAgH@e73nZڬ2Qi7U)1Ls6qiu3SX$B6B樈BȈZ:f9j-;s'A?? Ϣm4H*S슯ʮ eчynhHE&|0%b 9 ^ђk$֋Q0 Y$0ha"Bԓ]0MѓTw#@!(!JpO,U4d4X#8 [&6+6*'aI=LmMnH vh`#+ASgLٵ&(貀RAܘXʏ% Anm 9S̒%JvCWUh RV4VDfLd{4%N.Bט\l@M=Pc[GuG-t wPa-RH& 65c!qK]w͕(6`L˔!%R"3!&e!dJ,M)Z#d $.#(mԛw h5 0UGtۍ5z­XRl/YU|jTp " n;lbja7&qfy]XAS蠭ϘɓzL5AE@p2dFN^|"RB4YˏZCl J +EMNF. ~7RI( ׫%o 6H˕.ŏpΉwHMn[͇p z=֔Oџa*Ӫ?R`=+ؗ o{RUre^R֛ 1U!Kd,sdƁ1νUIǜ=ј<(Wi FD&7!ŰXUkYɁuZ '"4f,aX>14z6PLB$ĀۼE ^nvRtPJah.9|< [w do |)̄tTN1_DƬAbOm]Vg I\x+Q$ɣe o$\En{Vl~k](Љ|%MPdPY )鷕'?Ra([r['Z?ٻ' 5$.ƛDT߈J#k'i}!-RQfg/ɣ@ڼr%fxDޏ;\␎C:uU:WIb8?[qfh˱ b[%Cuӗӗ3&4]~8h%\-֡g5s~'5c=w2rd`fӳq~Ssk >9jZͷ /f[0ŮuDۺO<~+~@jYGЮ|T^: 񵕖N#˝OUzfϹo Z30 CPo5IyCeZc=-IxƼ! oR0!xeDEV&FbjPhKty0}Gs̚ WI;g( w;{rH@nuQ\<^JlKB_kfb֧W6A"+b///rviC:4 ;?{Owb-?@Pō2\1kʼ+R?Cn+/\/[jloB5cBuoƅANx*{"/ZE6|0Nj f]X`MzxR?^dsw]\<}t*Oc_xf=(@XπgbI)2#%pLQR`BNrSĔAy.}.5븍!AVj+6Dиw7@7~8SDA0z8@kܩo33ܩOwjFܩ_oYJ|$QpfA'5rxҫw}yEY1oJQR$U xeo3P+sՔ3AeHHAy%4=)5# ,$YG⸟jf,5ݡt3 UZ6ݺ>"]W1M&A zΐT-㒒J2Jt^fW];l{ƤіХ8]^\IA/5"TxHaFx~Iﵫ=ʧ*rÖ݉+:vIQdQMN턧VXY4uHM[-[a1*P"rIL$>R%ft@CP'v/CMwzj}܎|-0 "xs@84FDJ&f"֥̲D+Fw`KX s sP,G9 Q C(kFIj`4eHbPŤVkd8?/(K+ VAFS(\d5pIMQRKO>cȽo N;@5.)l- ` t\)Cd4,Nb%FБ_! 2{tQkHy1!㭂pWWQӨO>Ppݷ kmG~VljkmQ+/$)Ht`eJR4 {!X(IEI]6) kś] !טx(VMաƩѭDWi"L-p@3B:Z.|i5vsTGxywv'˳ GwU'xh/pQ &4vư$ XI4*FۘzcE5Uw7 _oAdUa/v9J)/Z#Gvz$gςAc}!<Bx=< ԥuw30yt CmO8\BFռ@-xݶ,bzmzg7!*5h~_BZtKIwA=pZ{ȨcaFK k׊SZR١G*7ppп~Xg&5b^zU /V9@L<;hh(Y`oO9_6$r <`f46,FU!Mߞ"uR)O` 2/7DAl6:lЁ7YC Lb9\6_٘{1}"K HFx&›ɷc­_㏞^.ȹ)pgoeb1no}!u'1꿸ߎAFlгx[i>F@}ӳʋ*f.jލٿdoo[ăO^+e4L2Qg&p!e?6}}R7sc>7޼z`|GIҼ˟lo~Zn~-23?3v<_uCox? cF& >h9!!z]*?c@TXCpDJG"JYR7ڷye]Xx坂d>6a7+IVOONj*5`Y <"7A~xk=B'D>iF D*uE4vV&J!bv8Ml1oZ ,ҥ X`Y=*JoD)íL IQ;t\hRFHc R̹)ӄ4I rnkyR܂TރEcvge0?0O }.aIzsƿsS[ |[^en`AP5ApjUC袈9l$Y[ɔbVS8hw7DDl OQ:"TFp\2Ԏi34NMR!'\2S4LnDgJmJ5FNYA`%A`BR~s;xe*?x9~U$^MvF7lΝiTI2N\v|QlS[?`u̒pͥo.{^9bl=a+DY~sً{O~ eK Gn)DmwϞO>zM&[.,ԙXqy[owF{=Z7M/_x'0mA֏~hGn*ORj['ԯ'5]x>wqbYp5 xAFhtea=n5rxl-yPZ׸ BZs&b+MNߺDH!G퐪 _GwV&wJJ;;L򓏆l"HzInʶW܇n払~s7C^ aA[Ѓ4hY <h3kpmAmJE~&[6"G/\. 8I4.v\|ɿ^ O}q*kT{pGsT_8\2k\(Ҫ_*0©XP'B1a1!$R#tJb:6.NM"3{7+1O bzF>8WM~x%7_|#ӻ7 B @aDKp!CVI%M@ݡnAn4$&nі\ Br9y=-rL;8*qB"?f HlU,V(30LB1_BX$WBCx۳fBuU^yU1M{x1g;U6.]iEưmdCɶnng3Yl!71_yNqɗ]Qf ͳܟ5A16ݧ#"Yh)olWܦC̆M@ &s.=Y@2HW.A2U(a(ξvSMhP1h"sBKk`Bj:$+ {= nv A vkn1-t,v &vCBr )[n\B8Fic(VۙZzgZ|""S8qP1hր@ۙZV!!_ɔҭ? nhʃ2:F`Jv &vCBr=[$buwf3(ƺ;33#/dvgi,pww`Μ}o}Yy>#skw+AkK:B[E 3)9|Dq cC6+/ۿTII9z[T0͌R wMջ?ݶ])޽ޥHH&QB ؈i K.cN0Jey]ugC6%O`uguK%`58dgf_l$a.6kƺ&݇ld煉rPbo_3]o}~W\t߾B7NF3n8_>GG4{C-Eň^jmgJS/xeF~L;»3a]Y-SiTll>GΗ -pƬف[w9 2BzY`d2*]eV/ecff 3Y.h*BsǎSqJNJq Byǩ848u%NivTTzSe^4?<%? C/gl]O= NhlއhWZ"=?Fpa#T9f_;7YEn&J<+>ϣX߇-W&lצymMJQ+"'55<.Nm] E^ 2E.5+%B u탮J$[-YG8%!V[Jt>|[|xJ4!D8P k6C&ROoi }&=fҒ̤}%Rc抙OY2'В~e-_%8io@溭EP]춲Iחme%}*,=Y ػ#Pf%Q&; Rq⸠/Z~9 vRRǰTň1e1W`KD>XS"M0!p#:BTD~V=8;9OrdkP~&³ĨKS NMx0@PF56W90fAc@L;ƺcGdhZ1[O*4-l^/I`H=l5·roV׻9PB)XiʤҘ I!Xa_ZY;Z!76Ua(u>|*P(R_ǻ$Pw烗kt]htZt忛)+1p @դ76G0{\\FM/݇dΨ0F`K !khypL*Nm-\ EG]b@I`e>Q}>Lc : Ha"@'FX0rFO};3%Κy-#.QvRdӖIwt hsŭ0W/in#Q: e|ʛty&1GC?Frd&#w~8IXLcg Kb?{OƱ_!e72T߇}xA8~X?sJxeHZ[=CMr(Ñİ%r"+iʲ0]'B}BE8 'aH gI!#d| c0YONA3 `[yEoY IF m=oyE_'d!ՠک<սҼaQDʏE\V  qEQ8U$X3>ڪDMз̷wbșwf2SFSE`#kF>{x?jp"*Uҩ2HXSU)"1KXebv\Θ$%BqI :e×"u[!Ake]RrCF&Ê'if! qBN|yD,B.c`n,{AZ|ȗM>pQc :+ͣ@2%H+$}_S9eeOӈ: I@xQub*!"F&J2b`"DŽ(G"4ЁX&E"HȐ+GGDH#0gXe/zB,^ >H!+ RSDBF8a_*c?jD>EK.`I9f2sX9VȖ9fOEfġ`jiD]J0V2w?^=:aDvf , _Qܳ7>vUM 1gA~<5bC`ð?v O~3&z fM3][jcd:v|ȐGެǘO2F]>4 wkH2LK{l!%B@%QRlTBZ}62M55BW1 TƧ8 *COKBONTϗv5t,7KR3%LȾZ +jt55-%l|,َL?Gn\VnS^Qjy4Y~͹)qX(I/NpSrlj{^YZ:QXb?>JbT|e aG>x@( bGL|M@0B]|%p2>MQ-]5r[X=]5g n~ O5&_M)D+*^FQh5ݩqZ"S<]vӋgk,{yrxW˕GU>3nCZ6ԚQûdGgӁNe &o}n8I0QDLEJ{6|9kVH[No~FooМ͚ 6DU}xs~؏)6I փXqD?ofs6@~,G)X^:q,dϮ2H+¬<8pϏҾ iY҈N?GˤŌaQtYWF@-`-fW6*p>:4g^'d5Ά  &9q*`!5nuέ|÷şt6cS2]"@eNGFSl?ރφI\Tdl[o}Jj*vn{' IFΧ#2ذ ##&ΡD"86t,x&?/AhEhlv)}e;~JR-M vXz:h%Op-,L83,Y:~6ky} kT9U|suOǣ`Me(1NЃhܝ(} $?q%ky}1VsEqC^J,zxAb'ZGi#ֆXpQ7)ߏ,.n_FgobT]}Hkw[mI/í5މL@W"?Ȓ!BQ(Eڔ0ymNdKw6kAuR +V2@TQ72pnpURDaFY2|PKw$,2O2Y_'GYrCї Q;p[&2@Rr:6WC=+w)a25=I]Ǐ JS7Jp/:)aw.R? @kivȆh |mҁMw '>5rAo=[\1'#2\ѭNӅT)"Z.[d ԇQ{A"ܤt n0G\;HdVXJ+Z1%2%q/B9C;JnM$ +w;),Ⱦ HPNIϖ퉊dr7Vsm; qŐM@s]9 >"RH@0~ P +ն nr,i&L(<M&O71܍TYïK¯fK W>K+/pKX,*͉dӏc5]c;ɻUϸ nL'Tݭ9e~+@c̉W_^ZTPIB5unQ[OwxYѽN7`EL(ޭYv |Ln _\' Gs.^~JqvntfQ]Zn'"`:*+σ' ĕPb]jnd5c!vqZKM9:H.$2x8ƄutĶ I.`MGLV#"ՆogQ%#.ˬKe*L( yO['tR~*$r I+?QRftUxj)ˈa)ZT΄t~i\M+lv\$r1#];Bh[;q$K.F^Wd 0a |FmxXKĆ}ҜvObUP]Q1OmVVx#F \Ill (GY7-+>f na:fW?~CJqᠯip(LX@&c2>2QGB`)D!8 6ற`4dh<6pE5b}Tf|iQbj ^3*9Pe#GC E!BaǾuRȏ0M Sy'JX$"wn'~.l50ڴ36-75dgT]'{6b~1s<ٺ _@~_3Z H&K БP'U=ƤAB$$:g]͘aZc:n`Gu'ǹ7Lէ(oIC*T^"]W5?waHњolrVmnp{A NɉJ݌}=W? Hn/>)?@;$!<G5W^"LOB&Z4U7ђKx)Q.QE1T$;^8*htrp٬nE9Y}|HD\"~lܩdUYa _)89k2A2<:AXⳲƜ亜ݙJ-e=usFs GsRr֞ݐ_)?.}WVRr0t;$f=-x@w~Ut\ ,i:K֯z"IGʯ]~ûX+Ϻs F3:RYr v6eJGsNj4"E 0B;aTF<Hj0 MP Z*.Qjԁ+DJA}s p*KAS8⎘u.*1RJgB 4c-Ks~ "U։NeU.e N]6;+л=!e)+喕~}y0SЬɖuÏ|XaG4J{дcGR,"*G1{s'v %8w4FΘvI2ɵ]Є-\[mK(搅Y7R.b|3Օvhk̚vk}֮èR0ԜW WvF4 /rrG1(uhk+[ڥe&I5n *BqGxHfeKP4y3K,Wy,J~;MI.<1k.P}@$c㹲bC-gqZ+mum@}Jamo*bso 1H-{+<n*H5==}0Tq4S=hF6*:[Re' RT;'*Na0:8,wtFѼ\SŦ{*1B]NGsrEűTX[]Xj͸;|z~kgK]^f4f,> Z`ߍFI~][zһޭnf6؎#,L(dHJZ$䑝3 Jd52ItɄREZl~Θq t9NM?[,=b fuN5?=` aZaJi?ݕu&^B-waJBQLN Mw,z⑒1UZ+u7djºM/}[CxH wc)p)5:bHm0qcׅW{5H:AHɝZSp8яς/E@Vz*$Z,΃/VH8kI74vN00ՁCOQhr'>Ķ ƇbFO à~m0'E(WXV-W8Rգ'c3:0Yj Ukv:.Pm,Np9uS1sW JF s3ou8Nv䯻w> 倱zŒlw7us*T1Mq.(T&M(1JQ*8|vm%$р;C$mjFPBK(gdW|M1`3yM3 hV.s9j`l.flIv(Qqvz/1߭_zX)CBo7'K^_ @8G'5lDEO_^{p6_r4wqC:7pp<߻dEskƥm|_T1ɥd9sC#!Qybz? akvUR4^{mh4ʹI=Bp1 ga82l_>uX֊, ,MrqXLWWY~\\ }>0-̴y\s. gq koaN &iM7MȎ G[g J a[Ѕw ܤ̴WgfVmdc3A`P#sDŒ pl}O/ XO`ku!{gز|#4|e u)krr N3Z!f7Z #.;j=C$Ҝv@|82WA5 \@R6&w +fϖXxӿήM)5n$.Ⱦes&x+\M4j= 6G&]%3jXROmzp ~17C*Hm6kH' m D}DDyS Vk[-~P> gl"ZMlS5u{6k?RhM5Bf,3VFG41_(Y{k44t͓#'  #xjМwBzc(`3Et>r>㒄/S{6gC%q*a d.o@Kx_T`k?muHpY` AɗcZ5-V[rL5R\ ?Hh&1SԄibRP+%R&#ELz1DVgpLSF,$i>cZjnLlw@ .~1d6`ydxu]o0N᷾-ʏ*,}`~& zVz;XԶ3`=̾~y nn2:%˖_\ͫ4 $%̇TYVR МH,}dp׵['d~gJJ)Y,S^D0B-I Yb 1ݦBj\os6Sr\K%\;!Tr9<`0VxıI0;"sZ:1vD(g<-INe!ݝ|y?:"օn4M:}T(7/tT& 3yn/GyP  XVpdf(I֧kohbl lOIٜA {Ko2NZ8]X`&bplOYZux-~IJl?>p2Y[+NZ2ru_e$xVAd@Wf9]liy% yhI̝!lӂ0FTMi2_[hIBf,'z;7/|i llGl?ObGW4uWsՠxٽzG˻8Un^oh O6b/x4Ϫ6Vʥȋy8xc= bfK(spS։m:Ww,'<6dܳ9J&a.!v2;gg-!օ)ϯctRđmd4Q7ڎcX؝DrTȭ0ry4N\^y_[\ baEMJv>NAjVo oQA ŘS>fTXPVhϣ*aYMLԅĹ>Es--&׭܄UvQ7Mﳟ-{)(j\ϊVVO]P~cA+/\=qtfE NuE RixhY vEO2:S:I\rW7=g8RE*`.,vDX<:&cBK+X;ܓr8N"v:X[,dB8XHY ~4Kl=\mn vXf[',ɪ|tfJDgJ؉ca9.y+-i-˜s~" 1vJ ѡQI(~c"NJFF`o2M U$)JUD!$$@za"$ʀa clDiHR5 (ͥV`0BM;TTp-Hc\HJDMHnoi\~f Ji-ֆMkL C]hARi_h v}%R͖s澵늉\`[}}@OC6`DJ'ҰKC,M"`2E$?*Fi-+m%V@[RCvEGa)mY,%04"ԑI)@EqpA?>jUة_01ݨ 3CvgRa24fBIC$> &Jbf炛8=UKQt{,-)G#j^ 56M Ifeh nQ)А@!c)KdfH֌ӄaS"fxRD#\XS"xx8T`bYZPǘs7A϶0GYv멝NyG3k7e V'[hl**j`l:Zv@jZ.#cdgѪ\>Ai&J YoXն~O,X],RAk bh~'d؄b'a3~NnsD. :5=‘#PR(i흽UIN;6GS#0Jh0TDN9`sE||a8IH"hہ0&D8Y$cp.7)UCJdHJ!fq;ڜHz^6F cX@0FzXh!a\̀pbXkDX鰷 ,b$LUd{u*yY|[2R!%B12a*&iLwM~ 7d 䘭|8mOuՀIsַ[G|Qv 0O{ŰC+yUY T3zsЃ /O%9s/'z< )"T  C.IIRC D" bRlN#WrT//YfmR~-dΊѴy}%X/SPL-s;kފּ3rb0Eg`hʀJ#iL5dXk!Pr8*;Jە @&>:9+C+1G{ @'1I+gO6˛&mSySQz_K]k_9Ͼ5\ihuiŗM7gvf?m1371/OjU!vogv-piȷD`=-Mi]9Fz71~=\'z\D]d XO9[H׾$9V õТ%2W?~=̚?bw偍*unߡӘ97ۻ-w{򕋨Lyj$)/0C3yǩ$Q0>*۟=Ņ~3J\`h]=1d45_J G"WS5ͷL)ݼ}N+p?̦juI!0P30~ Añt+ZAJ@: fNk*å3X/?Z0r{B B%=P Fi;I\H͐9o p JzMNgի_PlNT&YJ&eEhS{/YO@ҖhaªSƴEZx{ZӶ -M?E-W~䴅Im L)28T*c^iXu:6 D ïr} h+JZĪFA+s1 d{)O`Xy>Y{,9}XV(a1IIٲ5Srх-0Gm/4M%%4f$[5 p'>}agkaL(;>G9Ety:xsi]Y@;?S * jȺ\MƱDOK-D"?>_ܛ{s1ro.F#<`TB"b5 H(J$)UbiBBH)X(AZ@1LE8P. Δaʋ,Tf)XF2+6G_i8'R%Pqnaj^RRZe#EL#T0@IBJ$IGP'G[&f]Ȕ12)ӍMyz<-&V~cγw<5̧.;twEU-櫟Sg=]\_ X`P4'??\}J>M&z oLU7 i6ϛ M(@ۣ1bmh-KI9aBq!}JQmPw>IÌrAz̗fl%ʥax}q!\^Mf+cgZڌBϫ~Z61Ow*ve7WWiJ .գ '٣)>('3gۿh4O+S=_1< yځɄ"^ػ[6'﹞|(/pP5h^6zwb] & d2EL 8ZdRI.ch],ÂKL9<1T1j,x-?_0<)gA?}jYbfO6YJj[CR %m/2[>bMs=_m HS ' /~:}]m %\oqͱgȎqWYd)ufVL񒨻#c'zͼ- dW)Vmkx2_,a9X|,CìP{eTKrpOwQ("$6CP4MVܳaϗUf&x礖g?yHXrq0TE9HKUX691e|bfS@*T9t0^7Ul+HC nڠ ނ'hSQ38{8?? FAM0*V^h @(WpbLssH*z*ţUxm_[poŸ}aA4CG d>>a*\?5} ?>ejBkT6Ɇν5Gq/Go'O#^hߍ_ɓ~$0TJ> Tv TVj;t'hhmUeISKn2HQltDo[Wu֩VqZ>\ӭ/i+eXwFtt|] )m\=EEq֗a/`BRZc`.%v"OЎ"Q%=3ufIZz[aa?͑9ҝ9 ߵ5Iܝ6ne@Fػ~hQ̜u_#4х~(b}Ժˆ]bGV5bbb͟u%IMv־]fMBHam&hӠ'L2LX(%cCijkq˜IWOFQ$)NNagOVG~JT#*Th1#s.hY1(O_V.=QΊm.\ "<Ȁ0"EBA\g &>R,c>$VXo@ b=.01;a,@Ep(&MC.Lgyޯ˹2t*P䠞CU֌@ Vu-֛ GmFHh0?kl&ԧрh?sB+]Π,qN] 㑍4!&Uf@ZKZV4j>7[GɚFm۠l-P1JC-P@D P(.A(kZaJE!̦ D!y*P(%' Rm8! LQ8q<+,,1+'<Mhi/j9N?lelћiD3XiF($A׈b & _idal>DC)>Q&.LJÈk^J}H -GXd W ]nC#XR*h+gIc3Vw+Y8+ʷ3 @??Lѧ #c|HRl#iek4\6\,CkA] sݬƜ OYP 5?G7hDh` a6x&|Cu=X Pco9YPI^]oBaZmp4jY`cY ~ANcya@T~Q1 Pd11RL '=X(3,_X !3/q " ~Ba0\AM]rq3Mel/tJ7rrjȫ]8RFC0h#*XLZP-3F@#^c NqfmQC6̪_7ǏkS1G.|"E0\G[0X/ ާ?* Wd@Ji5Sk!%Z IDQp LJ@ťyiMvMͥ6!ZڍCuٍc4GPYxa#<Ҏf;- CS0xpz-t|cώ榰b);Jĕj]q=t 6\RBas=#xeD!8*HJ0_аV Xm)ZEn& &p8q2 WɌҕH͔~ʷ3BuӸ'Qش}[y-@lq6Ju6c?݊;nm C`L B/O^تVjc}Ļm:z02W hTtxM]|1R|ql`]{~0%\C PE:pQ?| 1'3a.3I= kN180=Fh D(oSͻ># i@Tp_+:*mt:LĒdb˰9,&.t+bqeVD$*ԟwZjZ>8J V@3˳˫}leL/mtr"[|^Ou> mtZ|^9Ou> mt;m缡JrYFJa.%a؁$#;-YS1DfJ"D ECU+ö4ӫBOT 456RI ]hGfLS$ ¦Kju^BpF > =UpNIG6 z=#&kbK$c6*H`h2D/(:.yVF@ǥoim'Qvq_ճHGKGVTq1TKMQ:ڦuɥC|z6kA}FRrsr߀'?Lo뻴BJ |77'*Og%YP>O;Wf6vy6;K瓀dqaߨ"/4Ò+Abu;IN<QL2bVJ]TCP'70uk.a,o{oo}ެ,v?51`ڻKUϰZ{!st L%iaX%ih|'˗ǧJj2eZbs5MpR YZ6Hɵ+bAT{5>MɃvNݩOfr4#x̰c| Iejki5d盗¾ b,s҇i6l3m_0 JBpv)dgPOk<BHq.sFԌPuMd/F+%DneUZ8 um$.74W;d"i'7צ 0f&9L+1()`(!j0d Sk ,xbi0ͧTm&WsBRČ@35B4{j4A % r ~9D}֪yFj爨ڜ.Lno&?j*?;uL}X-`Bdb,G1R X120A4m(}:C ť"nU:B{k%s;I=Ac%>evIf%%;!,A<ҌNS^vbF2ax ;TT:H,Z`BB'%EEV(AIۅlq1q p:9͌'>9RiQ0 AĠ' yeZb`nun$*fکzEx™ u|qx>֧%$]@+/;lv`E1JycOBbbH1 0QhT3<F1aeI (F0Pdw:̧7˼XJe=b( cFC\1"IS$r7t;挾ڝ9-3?J8Og@ȓ6#^y:'KkD8iׁו̈́f=B:ڱ䲇Q]xB^:+5vc+1V_XF^><ֆ7,j4_-n|X3,\' f+O~I+,8zm_\?fp\I,O>.%ӳSMȱ_QD7bN]1SV-]*n1Ztl)iИ؟z DS2D=ٛR!Bѕ;#Q_lvx7D]zFVœb*ժ\7]a巹W fߌ弣^0qJTj]|*diA9﨩[8\'H|Orj'd(DiL_~;E"Py.xwb{:eGmQy"ju r9tڵY?Q-i>eㅞ5@ l]<0-,Zfsmӽbâ@}:Aoe[ttn1`Q};K_]<{[#XT#6VrރMwt - گ7^f*e;M~< D,w|XrVz#)g g"LnJ6/;" :m4a㭞'֐5w4@3hç-vwDv+AtJhn=t )݊;nm CS@8yp4KW7Ku {Ka"D$8C ^QD9(G(wRc ~ r(W+6wB T, ۷"VA:"ٛW"f։ ZXFAb…3fjr#q7Jioa|SjB9YY&dZӵYhgLZP}-v Ju6dή٭B9D0%=-vc b T9)m 3`DحB9DK0EGiO.5\ &(KF,UG a%52a)4` ERs3 Zyө0MBt0Ϭ eE\r ݁ p˸1CBCW8{D\ bT9RFlZ1z+?9gLmLnjBde(xؔw4;aڭMXm>;mݺ΢E<7<)nܳ%( Zsbh Q+P$Xp"PTN#y*6`v:RrRBP:X)&hs3B(mnJ|9Үa41(AR6oRlg![Cp-).ޒ,i7EAT0 >vۣw4Mݶv+vC^8񔀃8nnY!])|Fe{"gnYYtc2(Xr=wn~.TR0J}nNj-%w85m}`Ry܏KpOI6Uxr+Bԝ?cGƽ,tJvChBt;(?e ,֚t@0f[DJE!] NrJ6)(s V6fj#ا",[bdx= S- ؗ<L rJ!хQ7ysV]p\paJ }xaJ{7D}NID+](.^Eept)%%ҥ%}G)HIb.@79R *D>DhH{u}qk94VY3V⹩&(, d6tvfgy!$qk.{o?׸͖&'hqK=K8tzrV5Q6,op2qHukkܒ߽Ov~h֔CB%r9;ӌq2P+YyGoY!\NΪ37?C ܕ>x>9&+͟]<(WoBSKotZUΜq׀oBk]k7YaYސq)Eww:L]ϝ?!x=żun8Y7k௷INrtuzt5_OFh%3Q@H@EPM"05^x?{ 6]/C77riuuANAozk؂K9uY\b: / C8'kYvW`ݾf/;xNsq&A}fTь V+i@A( J,i(*2b{ UheǤv5z5BrnTQgA=E.ɥ**[kj5L'Q|%6vxKU66~_h#&pZ'\v Vkm2*5yF}v)$Ԥ16*f-KJN ANl$I'k҇eaKJd+qI'u+Q/`)idHEuR$‚FŪDe ߸AE_ِP)| W(@=sNedtb*F@K: t_>z gP|йWq'ܯgؔq _9QNID$7HB!il,HG%DR 9{ky<(ODkVT4H#@l߲ bdq%Y,- Uw՝~=r5hVm=ɚh?䮛1erOwzrBVr" jZnE>f`q9:$epbMʋ)0.PUGJ<s%j6oy"O`(R.r W@ly E^ֲ5hǤ0\Dy=9֔*W-P 9O k[<1ANB+dhVit]DB.Eo`Ir@HsdgCm7~nw)ɬf}.n^C F!,`-@9;-@ h ۍfV)!b`(;!f4h ܞ$/GH:)9ZC5iwx#6*>Чz'xջw^tE=akWKkcP Z%h&*'?W-.qt.ý[;Ϸ"tl<8:ϸTbd~AMǘ14C-o&wy3ޯej裙D #;MU'$EӻFIL7י*\ _ ,)ܻ Hy TU[#X+Q5F//<0PjTZ4vWȐq'D 7oULCgVV&4,ӋV>d)F@> Bc]Ɛ!gk7ַqژ}gkӺ^Q%r I% FRD*Zbذ_];u!o8M 6_8>-6 9e%=&T@|:]_~YT A(ѫ]ڱT:vȺt񣶧f/j'*sLI/#rrW]sb3`UKi7:+ݕ +4J 5/ٽz=C8׳3Ȣ 'ZfKwNou7qy@%49tanjM/Ko3BC[ZNuNI=9/EεVĊ>#FX"8̭r:[Ȁ,^v](;}$]j!1vso&zM)J@at^T9->0 FQǬgP^$VgHr7}L L=իwoVj|7r]\Y-d })~>=yq;al?mcs;|ۏ fYN1GvTo~']TF'͞ޔܽU4C)z1867kAq(C/ A(HcekW#l n?fVWo(:~RFK9FTp?&,pҤ Ө0`#і9Fnp/ޞ\V~,WF}е|n/{snfOٷ}4x89|rS ptmhdJ9dFG۰Q< -'o 2D>ڄ'\$<`·B ? Oe rH,#J}B) pRB)#r 7z;1Io( p(A bBȀD3N rV``{(R7 HL QՅwC[\–M&x+AGQ߿tD|2Z}Us7ύdߋfZ'HgQKS#p:X6ʚ-sF5ۤh m:wr,SڰTGVm͜@WUZڎKuNqqS/ˠo A5w;2J!@C*| 86HPK%% 2Rnh\+`q)Sm+ʡ͞%Mf ?"SWW;tڗnxﴕ sJF~cx`M)e jLݪԮ2y {-S'aV\Ozlά3b Wmtl[Q7ϻSSdV1@6]tV3.ZEY<%\#` ~·Yq:.E!S ':DG3sx#xO;#r 8TQh EGAIr1""Hi*iYz|4Oef y}8/Wy fK-r@(e03)5L(Ղx}ႛ yPٕ#m^붽;1~fit6aQcyJha(̇)#CDJRWn$ATLf0+t.2Ε.'9g,Qv*J|] "gf" (7Z(1 ʌ02,3H8#F (%*iDCQ3J@T)(@&F[1Bb2Tf"9Mpd5 uሺBjǧi̐ a2OWk޳_~/#曇}#[mP$eNQ%fCme&x{ö x R z 0..]?-.S9VB+L)03 bҡH/ "^#!HH_=1eZONH5ى RV& *B#2f SĤQSlVFĜ\^ko=zMa19޶n_n[.{J#f4b6;Rv%JJ#A6*6^T}QƬ0IR!70 /! ^`?Do");caN5üU߽ ~~kW];~ޚ~%eJU#Vc矯.聟 ua \(?4e噂 d˞fݗc;5;hbGx xU*rfGr*\VA7jK#1Y۔vk\Wc,[j D3'g9-׹9@/RCM+TCC^vE׆Z4U1~[a(ޱA^fy~I݃8 hrb_A,\Ox (OZ7MMl˲zcj UZyC;R,?*q~T taqx(Տ|?I:0_ tס^*.Lί? Y.=g*Ø dtNg~|kyR(ʬHRkǷqwW,'ܪ|Rvȕ==5m rMqQe/j4̔G,`dOdFwzoڤj a6* N_FC0sQ_fv g\v/* {0vιDPUs@U'0 0<9xZsO; d}il@k0CfH[Hz/a<*սGQALƺGbRJ-)o^o^J)=}G2HϭC2ZCEk[={ƹVЩ$OR() /0|a&~ IKc /eB" R>urN/zJ;V.B%P+zMqTyl{Q2tHF~W/?,S^ L!|*0Eh*>%kyQ#;rgBC*a%cc?eߘ (L0jy0LUTvIU.(Po㈼2~J00m(k9% S& vٛmڽrZ_P~s ȉH%D$0%2KJA \j5 dl\|؝W G!ߝɢJ0b'59Ү{X'"[&G*9}qO 0]NԴq}9$1Ui ١>|Q:':*\#ޯra'Jjr\%'vJr_A^%]5/鶾|^%/B`呻Fo9nG箒(;!}kNܜͼv׍O {'z_%AvJŹ\ ͥwH=;mb4z5"!fx* ;ooT0.zdqzSبpftM4*!=%QGtG6 N牉I0IQ p $G+ն1âa\&9LWA.:l|::ˍAzrϓĄ~|@1)7RWNHcI9QS0_W G:6CI?>پcqm^R@LJ硗Sh9?B9j(2~<6CAT th37ǒ 4MQR$ڂG#<tԖc%F=w0xZݼA4&O " Wk20A@m@4\k6xK EZ9`Dj9.MoJ~.Y|Ԟ{h6rBg3p#cE |f4&8 RGB?C(hm˂VFDp`Tk# s R20t6um|I4яz_`7Qt=)}1Vf\"1?>Q|k)Zo+0t JaQмέ[%B` .OuVd9!ŅdXeƥ&|X)3N>VbȏM|z;Ĉe>̤fdY5*,ndvH-#3A&Pabbe~9I5won}c>,ܕNOqm{XAZhV^xŠx+܍oy&F2cA`N5h 0:`4BD+##В䙉!KIA>Q$E%tE;@m5(MX]"3,Z4Y:XS6`$韎1(:xLYevPD1F#"{> G6%g“"dREBB ]J-%C3)3ɮ^pXv$r$"&q<]Ѷ#tn)!Tc UUȂE=@EdQ=sofU{w6ϭN+xɕebr{q_ѤTz9/ǟbŻW Q$hdpE{H^k1h T"q}2C>@ G\㊶"!2\@QvDGR+O)32&:{"h=d :ǎ$' # Gp1:Ĺ /#==ͧC8RG">~R#(X`US% *Κ4!jF݌K.}"׮}$ԐVQ2Z+{'yJSRC K)L"y *mGILpe*Bg"3S3!Q8J7+PHʆ$COJ<ä1(& 'O: 3VĄϴbCPHC $8RxZ* |EDGBErɑ(ˆRC$؃^56R [,e~zzOzFfu0/ĻL_|r`tAn,Zz6Xt@{8E͢D[U;2%.ʶ5S|'ĥp: i0Lf]G4B1b&Po y,NB~?cvaQ!~Yw!=^Q\pLjMvɧ0?%&:3F+'Bw\GZoWV"Ÿ́1?cDs3_ ‡UA#͑PJ !76[]k&cCzrJy)e񭗽+l7?)2g+``Ԛ) Z.њNS dr cIfm:$f(YR̕,peY^M\%\6O"Yֻ \^kS dp`u9r t4ܰmree Nrd%ϐ2h!S,w!d" ägwCS Kuw9#92i^B=ƠLVy5Uꚦz7~SΦ[ubTC eE o'h%gEXl0Tp@7+ՁY`"OG=H\13ydmH`;Ww6xvzGCN6P#Ӟ?iq3_ZE_+J}E< "7Ϙ>sW0dj ΢'E:wDB>[-KD"ީhGGi1&d;>).T5H*O~SpzQ򺶹Q#=KWǧQ$[oß^y*ofi2 [9}aU861_~g UW~W9jQk-BWbEHY.{'饧uAq[*, Od׉c1cT,0$xyc2s4&Z#*%nq䣶8QCkt! n`5+!rRFG>_]rBݑvݚ, IiCW׫Pøիd⭻;&pG-*,{/_np2̯+n9'^lN;\\#yn_;s}.ٍAw!/,Nj43= #+FXy,Y텧߷rW2Nޮ mSNiKOz}X9ߑ|"Li]n:n|B.7lQzw֎7J4 Q nGP[CMKam(:xn[!`]pzw‚EwG(4LRƤmQP {CD%62:XVEnNQBJM59"̉QNYwpv9&7s\%ڦ˓C$ua5{iIgҧy?Hϥk)g|āpSlC;aҎĮ6+?G@bpdkN.g_7Ϭ]e:` g٥vlHuU]>"-8g{(~\3;mv=~ {n&yr][X\!|(`J?o2uq?jpVନYQԧ.Bf^gnj VzւB' /7Tx.h+,SŮ/ZObW_z6.㞀ĹMZ$``l1Ї1hiX:W.Ut 9Z߮I${#AR(w_VTxl_t7{kuAk%plo_^0Y\^ЎBkcՇW])n=d ;[nQ%(L=:0{E9gWcoV~+Γ\+bjX77:/?m~c:۹+vy!WC ,pS ;X4!㏫'az ^{mF-sr7vޢCy"ո#8[½5S|,bY7<w:jw uG9ޝM6)`T묛?|'"j hj^sFfdFDVFi4&KҔ,,HyEfBx_Mda47!`VE0I XK!+es!3*HO8 eԼ[s'J 7/mñF4I(]!A3tAe["[h4qNaIq)eQVQ%+YΘ"$EiC8 k4RqʬJi(f -y<#|L˱+_f_nEr?Mj؟q߉q FKl g@@۝(f4j^{UsϛB [#_64v̜;?G>|y,܈Q][s[7+,L w4T}q[SM%[*\mŲ('muH"υD%u9:ht7/.Np;˙{\>aq4A}ţPF=} {Ds@*z>S8 8LHٍmtT A%q5|)rI4 O <&@mFDvBSABtQbF2^hl1[OC6*4ztv7(Q9Cd-Ep gw8Vi8ELW9AI2`d 4U,DNIGd.-zE )HUQEk|ca c̈́f_ 7.FHΧ/Đ5ZI/c5Ҩ XbJB,Ȧ+PdR$9NDƠK Jճv{tu'lQR2b֟PF68drJ9y!b.P"LQ)$:=@HAD Fr:4n JԘA敒*4^QyaDc٦@t6Vϖ> l)5}ԱJ P}b0Ub $bớڢP$E 1 69A($eڳb?B@c[ 隽ҿǦP[T m:W:Q|cnK)i'Pz_O !jJ ~0`MLyʫ mu[]fwoP(]VG-z5r ;E.>;'V**vqcrUlq'ϠATI=,2Z39Sk(o1@qPSset%ƯRҨegy9K|>eãyIpq;6rt'i@:(J3q,l0!9Wo;Tr G L {1zjN@.29m4>VOߛyM`NFj<5ՎDяNdXRҩofVfRkSQC$}G]G40Ae'>i[F P0S99.{}JB4Q.ẃT%3Y^Eas|!ĤjCz kDI#[ g%zw(xcA%(e͸gJQB+$ C]Pv_4nĊی ~@_m>#ա.9ʍW)u@OAlP-˃s4d[~De h-7h ?!B.HyOIfpΛ'ϟX"3q3 ,}'矑ьhe&˸~yo9g<[r>[΋u9q*#E!-R"!Ip`!yd>&'h?#<}?\ϢBr]ۆFgwG\/;agc\ ޻y*km&/ RM0*{|}oU>i6 p6+ĆVHGɭQ΁{RZqsX8 YAEO 6Hkmk5ְ+^>mPQQfI:l :T ^^G_7@Hm3k^Q[#ݽ\IA$RO/=0c3o<*v>ogCe2lT݌'7ggl(^Ak;9pO3E퓝$\׮efq3^;5m$u6İ" EF2Ƨvf5Fki%g*P₤\`u:FW*_@()$f2']GM)9 TRzt<RsD8 D4ѸK66EP 5QVy1!s;uͱ2zWka*FcR4߆./%XoAd(y~ Pn!mABP^ 0̠|o.X>h *(yxSPݮb ,i>`t΢8gFW=MU8XvYT=X7<_V_pj)Ǫ5 h6/G "8Tކ.79|d&r x;A (+N?fClD*?LOۛl7խ']"g@=s?ً˻I]Mooд"ӛ41h…Mh%|'(sfl4:U:5?3wYtѢAv49`C4sETBL}j)]" fNfg.x)rK(F1EDP+'!D LytQg/pBzmj!W@U"[̏)pⶸ2dQw^{uUtٵY5%Vzツr8^$ܬj}ujS~ZԪZ$[ IRDipbn8L<??(e P}{TSŇϳPC{ [PۊM0P_3Y+^C~?_e~#+fVKb󮯡_$-Q/ފ N! b - -r[">IbU;= b +}mŞе;g٬>٦"m/vbA@J , Mb* 4TB_dE*P]6BW(iakRo"r$z'OͮǹEmh%e2J5.03Phȍ&Z׻#'}rJ9K[1&Tw9>ьi^>I-zDQ;\PX/#f>[:s&/|IB$=ew\1J7׶̡R_)=]DHQT%Te 5p0.5re %(kSBkˉ%IYxIR1*urI.H"HeR9\_$.FK 㨏27_ldYΤm\FC(SZEs|Y]i|z=n3kũxc JTǾN>mz4Jmj1, ;K9P[]pb74In|@6kU΁>}&&k~wFAlE"{t?-öרNWQK`Tā'6svF9p`6V)d|9byV˒k>w $SRO]{XƫV X\^PvJhNµgL(I%uzp& Q]W>rP(3dΣp|%I ? GQ(nJ@-!?.Σ0:Q98 &J(yZi}֟goy֟goy[_/, u"Fj#a T>׉O*'FL8-,.,޶V6V,?&T;,k%gR%V΋G7 ERmg2%[a >w޺TnR*l AX$`|d&ei<#Pr_Bj2挳v &[+_51:q>xXZSW/Pk4W/Đ)PIZڅmRFtWg;*+JTiyC_Eg˝t9>sq.fyWu;0ލ t{axIX].K0 ߰!7BU= |#(ksBX.iyվUW_#Ӥ~= ǰؒ7k} nV9G.Ts^ht8'9DPVmN {yEzOc6=s9zS,iC/,#Ag,"hMڪKJv@Jp*@R%C%vn}8)IšVfe"'2Ned1&SN"!Ip`!yR`Lhjo=wUyPP[Oh^KgQzpg)uύQZ߾c]79ق-N4BhodnӃO1 UzT &sA%\JTըMDԠ'(E2Nh%}ܽה)M`& Tum,1|Eד_OځvpٕAp`ʊƳQF>7tZ; cڈxp j~=Dr6,*6tv)$v˨&V8q#hPQ`$Y$'T#zF.}\dAJҠ9"fpNH!G$P )FrjCe1dak8 /Nwm=nJzs6;Kqy9dB&OŖ4ӒZ3lE[[76UXw: 1cHJUY9˵ʵDùV5b%MԳ\-%D>7 %ڀҌ2(/r5Nϵ!瞆qBNy݉ PBui$->j F64pdoB1 1> d͚Ğd>c&?dH/}U&U߬:c"N~5{!0vWq.~*ԮnχO|g7$^׋!ʔ|Iߨ1њR=:i{MS+ P\pBJdf8Yb5U1\AH0 g(7c3l4Ism՘"zT)뵡fVq#sδ@FQ i n{pc쨀$lXqƤd7jmMasp##(5N\ܸ¶rLUeYQӄaRa9R},;H $WnsF@r`P#l 4GOH91)EYV#WV)4ݛ#O Us..{7ſ#C.8uv|O;{u9^-*ng\|bm6xdCߝ{Rx׷k,$魪]>NOM{3rI1;$\QpƎ.!MY&}r:Ԓ&oNiUP )?0-?*vN?.(g^Qg Zx5<`+(Ew-4M>nF /%sh갴ȣ|x)<3F ct$2# >){mwf/K -[P3&PPs% N,7H>xK̬1k-=q,ի{u@= T{vi gYn5rEW]w 5-SdF/BpA4P+Ix9(:&Y.27u!6e뾯xgMҬ$4$v$ٱ 3Ƒ4Ӻgk՛.;DHZX'8Y2K̡`(V]zh&bj@s' Ê_/]rVB#ԾoCn4kſFe=e-r?kNS=n\`R\G]Н:tNM GuL}#P2hbu?B鱽U=L#HSˤ/GPZB/h)EPÏl5#菊ԠʘH׋ AKL)ZOU qw:jG7Z>?Ģ(ǚPI14T!e]1 L"D/+:P}Db170w@=ODn'+U%9Y|g-0j!ka(B\ǜ"/~~6nmґYZ)lBY GPAP;Uy}*I:2|@*HRj Iz*uK.y $1gIGbIG*%w,m/%J6WW$)yQxJ$- /$J;u8٧2ey$L#RVspHnC[15OTLhMA9+7ChkƔsk{)_J^D-\1p/ \ nUm,`j dB}6 1k)F.:OĹ73:.?1`ڔ2Jݻ gz0!+/̦:yg]_ AZq:zF$`1ޗk N-< !G̗)3)T(?m*p \47Z`vR V#O! Aw̧Ќ)ذ >]r) :6#O1%[ I K[T֠=rwխ[N:#LtI{X>]`OgZw!N `.9>)h 4J0JZpڶ999݀]~ 4?*EdR3)exȽ`jUT~j!glv9V\gfgא 9H0F6׸(ViKeXmY&{M"% ~)(b`krzD %[:])H*G]ۍ11A3%JYb/hhm'tqH萏@{7|d oxw %S |'=A~S=vO/0LS֗% Ҩ(St2l%}v[y s'_o i ,J n/|`GG.=[}@`p}3_h hgk\tumhx[ÿ!JR~W\R|pY`'uAƥJ`X\u4\su+[:#HbwS^hIݟ*]s5'1]s5'tn6ڳ,* dˣL)ܸF[؋|Z٥6GMt6X" |?iFC|`!lVߟ5!+#$NߌSgLNĩ}$ ;?p%b'QND;)>qj $ujꅵs-MVLj%#rAFAFPræj>|oFf4Ȓ ?lˍY-~}^ʁ;vyԔM;;q ?7@ZF0^G3$!ͦB4ױOڜR ZƲtAXJ͸ 4HSi}-P$$i/{N,G srp`%pq☳ a2Iqfs65ZZg9uo<ܣ? ph%ܸpk:@xNA`dS&jصyej޻ /J<mhty+EV@X,s,2ܭ ߤy A-]xC? {xy>cNS+#\Fjs"`Zr#> N4@ZyڵULN#BiC`h4`9a,^B9e4'Z r$τoBi ln5%f'l[{$ 1aq"8DD"N"'e":@482 ;s+S,ϤwRNkP('NjDn LQKZ#lp]?Rׄ#L)+W̐cQuhAA:-W\Yk2 ^[Ít۬Y`֔1Sg |cQUzD8>P,(uQ&j:iGQwfc>s5,XukwC5֛7 yǷ>9yCH);Y̶ki2xUx!_oVn=P߬F9cT>^|Ho m]U -zxs{~/xCp l{^ nү{8iH1>D*U_q (pJ "W(=ojjL lrL: K<` ExFe 'y(.hEz.r2n[d4Ӗ12gT9 g٧I5P܅oe+Ÿ,!B,$Q i(dPT5!L;֝0Nf}r LtV5]T+RJBi V*UXB=B)NCEA0xᒦT 14BXLJU~(M6hc #VeT TK#/d^oi:X=2WGێ㕕N2{(vY%JR IR!+޿2hm44|*7X+4%QRjZ2L+; H(ߠ@jȟF@XBٝіn|3pa6V5gȌO]ڧ74_˗9j$ǟq /A {uì\8;>WV@I (aBLRf9`'b)>P#Μ4Z&5hF{U$oڎc+@U2P+@U?:Fc@t fDtKts`߇'bJLQ8kc&$EPK!|NQAtbj~0MPǫ:]9dzlϲq>,ϲv^.2. Dy6qx` %ЖV@%g"ځj_Zhɨ3`4[Ёk['{5ףO#The+þciabf=(FDIF5^DV>5LiJOs7't?|Nw,Ra̧Pd8*DٶXKLpћZ57ys7-F[W14k+&!x27s[蒁vFWrAN@ћ mцwSdM S)(Ɔ@)_(b!nfNR\NwEEq`*B$Y!_f8Z,k w'@"f,fi  ;4z#&=KlCdhi AQ^ rf?h=WvCToVzs^Y}ް3͒qzűpMF8ܛŶrlj$+',@)^zj+PWȁu}( tcAGaUqVYeU|i.9֞)W$܁6Lh@u|+ߙQ^>bιޛy90:WlRa6?}8idCC5;O/w>~l"O8]}ƄO/; )4Eqtww6Տn&wl@Ϳ\-o7{ ܝ9&D˽uDݼmf"aqo(9?{89o Qy'8!W5U JQrF\-1`LHH*X1e.5B,}OscJIm(dHF[|2A34O@B! 5!+(84SV{`LufnuպSWjAŬQW߁! V<7f,Q j>]+ߍȗ5|ZJrCi6B" jP_1RSfN6ܬ2\G( e,gp7T=koL'@v|KHw=]E^O//Tfh >^>x&Ӯɂm=qˡBH*/^e-]xscy|=+GG{;{u-Ä%gDeKOªF0ZX819) h4 \-d,1Zp$#VU.(nlLJQuFf@]4Q4 &̈B B `FhJg'G馩F Pgbۢ St}rDs(gvKφWr0D Pb11'L%8 2lbOu5>;HLrWKѥŝ0)ikHH**Z&"όVi{p;}ctb{qvˈ\NA0wu+վZ0^.n,i̺0Qnqd0D\X bѶ|9pL4oK!L/etϞ@# h Uřg.EYl Q=u (Y6X"~;LCn~qxfFUaR3 02:{EK MɀH5/MmB/XR\pO̪K[ z %)?[Y@ɓZ6o^M'/2\YA N-\⇅2X?gWars#Z{6U Ty&|dzNw; v'N_~/5P'ZĈ_2DzѯeML&g\t߀Dޣ!,^V e+AdW*e:Y-ƗgTy¯^b\ޡ乍{я/MZF ˹< sG1PY>ioTCґ" )CN}ߦ{{яX=7iHQ %?U?W:\t& rN~ y6A:e4 Ki+o@@L:@T\<W'TU\;/KBN-ǿL(9=K1j^MYIt& >0NP$%QGftPRxC8wH&g1YH.f}E /F|Da#E4P9,.) DS,'O߬%<tj T U#jIUtOI;Z|d4^3C,$J"Q[a<&2)vBtT<#|պA`΅ժ_eLFG(Uajky:.53_ӗAQ*g7P{!ZĆ[,S&藝p g9Z<-7]~73Z<=8 7OcO͌:QTKAlj+?^QpnH{M,u״ł(M{]wT! )Ll};w,ZMsib2!x}-Lӫ]+U16[Rf Ԫ]6Tdt%]oekxJsnQh1*E]:кP/Ŭ vwÓm;5^~s;JCҤz로<zJ`+k Qgo|p*=gceJ a`z k6Ü'*7Qh@טD|õ6 Dn^wMPot*ۍNHIepD6n}sgJOݺwR齞upvH^-CU,]h3--/OI.g).[ Sr!ZD>=r@E.CHZ~j.4BˆjZ-+_GH]0Pڴ /YJ+^/CI$J<(F#&g1!4i І hM-mRgAvFV!/6CfJlҍRQ ǾQ4 Q.`1"vN~@űerG~*i 'R,3^N{:.B*?^'3'rVIm\$7;րcqTFX>rC*֢M;ӆFK"Jc(xaDCyU%䏛 J񫛕sg'KmN!WdStf۠LqvVJRGkz0N{ _VwB;,?{u|ynoD̀I(mgؑ٪O%ԑG\ED@]B}[|rE̾ \EF1܏6T<318aGy;'_E&WǙ.svL n?݋G9t 3%ubk"1UG2y.Izzc̱?L%gnat.EzQD9aT  >Fm,;Ƃ~͇1|5TGjTVKAOH}dq34t!cnv (ggiGךcw hO*0Yż?^+{UJy=|LqNq;AU'yJ: %1́+Gcr'.-Ȩ<9O$Ah`@U{vSQY,yyYNPrYzYPIcR*Z8O:Ұʆ}D.!3&wmH_aij%n.{I屟r E)$۹CJ, gTeG t%Xh[A!NCyID % 9`H ҹ17(֧mSђ_[_晕\))mG>yWt']k_PGgԩ>h &3'5"Q|x֚gf4ρTYvΨ8e^L[UPH = 2pHIt2. }|,9xӑxw/JBY'A-4?Qֳ% 2he>=5BN"Yl΄]zs:D b'= 4շ U\Q׿ nuj< Z}<;-vv@^\=:2m߷` +ZN|Bm 3VbA)(PJƇlK︷YֻPe6a,!tQ[9dVw\Y3^0ZCLaWYzPdV*`48VF>npD5Mˠ;YR$e2%eF d F8R)ۃ}7O |$||Z\"[sϭ>KvJ9ևm堅fϷ>|_m1+H;J! rhCHR lxfsT?#L=uγĺf.yTBB˞2㼠Fm~*r^ -ӧM+e zSYR0%C^83Q,9EJĪ 4Ewv2J.xhux f%A !枙ª%a'N/'ABGSdYeQwT/8mFི0i!i'cв: h eЖyD[&3(3D i@2 :4WWΚHະ\iEEׯ_ͪ[ZOq&W|R+>!IZUlOD7 ]mO˰FY͵Q,hS}Z{?ݨ<)7rS}!efxK-aP{r Z 5F/bԞPm-T&JO+1(QT5C`aP)M3*+~Jp*S/&ҐP 0`GiA{|t'2m"WfKQi/DO|4 BwFNNG8i(q]jDJÄ@&2w|*d#=FhI4lBv);2KCh+P<@@r;OvIIf/T;\"h#{W;$KnqIn|f ,9bq˔5YQLP6ջx? ->I@ZuXrw#)abd#1H@D$+}4J.6V{t=T0& -q.3- h+_ܶ.9sE ՅKtpLsx!= JpA qgق"j_ƴXT`Qަ_ (]4-%&DeQZE3r˲75tv7]Ѐy'xTW}#q ?U1|Ĥl}7/f mElikYgO("A>XUhRDbJE4a("c@3F 1uGK~ͮb ູ"ݯ4Q}ciG\qx(qX]DFhy:k6)[OvC /6ۚB̆h6)Mx}{|Ն}IOPKն;yWtRRG 7YV]R7gr`nsBI^7w0ެZ>'[cmKH_Aw,)O L7q^}D|q1+To({]ޚZ1sovEU&\ԞrԑUx2#M.{~^?*ECNNz2~ntZ *ь|/_+3݁%ye4z:ݦww=\`ÿ䎮<^ %*J" ,7S2'jV7_D8zEۥ+:~5. n7^]8Yw:VՋd\XE#KC_޽p* J-n4&SmS.8"Z yF[Юiv 筈CtI*կnpF -t'ܬ|yL^`D D4_ou#`[4S$JETY %ZUገpZNbɤKIn>Лml{[;UNraA&h>Kl&tFB@&uD3:^/wQC( bu..Y2"f.R1}1FڣI-zPnh[mSYl^NhѶyB I|,osE`Ѻb赃 "+!ӅfѳZYfJchm2V vgnU3FOA,2G,Ԏ4-= QS8%sC>Š4qv$Qя} aKɓJ+l^[Y0żD kepRҩNmLRqZ@ VR5+M!MfVEv"kY>_;߾(,lnCiBAEDS/$+bID䄈мL@Pc1n(+s>^](CzH~NS}̜n!hg~~=9VrZt GVX@Ev^硒*%a<[)}~UT}Fƫ׻эӇAǬ>&em7wۻū|ix7yJ^j2@fz3{:@Ae:P}};ZP0U[vW  M:Xo]szOK"jsP!!vZ@~ Ur }v: ͜V_wo7\|{꽟kc 0nkIn7:|d7Z^}r'\ %>2]r zg%jyM5Ƕo|'g%sCy&) %?If8Z}sջ#tӿ|ܵaHoFeõʊn6B}p\':.먿DRNY 1itA3l7߬o/r Zjľ=v.\y~ys˺C2&"̩.r,>0U%_(sKC8F S@CH2kR16gfSN6;[]t ʅQjOЅ-rctIdam1g23#n&ć['cKopǯNjke]ڤvoY22n2i%iT&'{Vj 9oH. S`΁sV"݁Z|W F̅1Dh5^ ˒ W V(-XBH롏a>֝Y 6Pі-Y&no6*N^(.'&0d[? dR UzN 7zxzٸւ9/~c)B&8Q:f(z.}Y23Q3Hqq+ xAj`¦kF-<xaMiy9h ߌ؏HWj&$Pe 0FAH`-g!J2vBJ#E-qJޮגF 3y/kZ08rm¿M׎S ǥj֕|=io#7E/,dH:b23A|I`vK#ݝttY,![iw;FldnJR{-Z鿼'WW!Z+bs?we'L(RJْ,_7A2^?7p*>Hr  03f@ɬE82Sq9|$Kl҂%16;>K1"s N ./CNP\ -J9(Q:Xb*u[ˤn})lotDU?ZPLOF{XJwP ̀k!9v[P Ƙf{<9dbGraZ{QY+{-%z*-x]`H`>I-(fdFDtvWdg/]0'FT%O`Y{-[3qQξKFFcE/ykzTFGNYДQbpVʹ n*(Ԋ*9Ѥw2\0HKoB69Z!rڔ2тlH '->,\qShiA`*C#83;J*a4x6G6C_h KeRNv딏߿K?MNx?lXLjx*29"a$2ibQ }6z_?h>D{`= %gz-_(=;=6}WkkH((9!]#R!HW}$.ӈg/zdS1?F U'N k}q$+Y3b&mb>p/" !-ohqw Z}s?~F$Y?M&1Fpe{?i?/P'Q ;W%_Շ)*RF J^67Z[s*A:‰/#8w~y׾0T/? t-a Bf(ͥP]x$#DЖKB_dԳDWF3EXen 27XRx-% %Lh^R&}8Ӂy DΤ X\(7)IcIǞx뵸\g)NiՀ؊hCzw[CnͧF@:I y /IPHD'2fF̌"ȭ6s4哨j̈́hhä!=]'^(t@뎝 eC -ӡ\+'V֊sJ4JuMYoM4L|E JF/}pv7ra2S~~V z&2MdЛ*.p8+ +.R/< {8q; 8[k ͏%oqo!4zcY!aP3PA(gCf˼s0 I?淃o< ""i~]V۾ OlOɐ ze(_MgƊ1:b]{[>y`[8ϥ?`4BHݫZiM߅O72r-T(ف9[%*+V\'[ hh_q 1؎!D eNc*ZT#OS d~V54+TPʨ8N?ȱ;6 Uo-pvY"t-xmOGbr6iвMqémhM1OJ&oSfדr@d+Ox.X;;եB* uJs9|ͿB̼&MPP L;~4{_IG$$oW˹  R8tW\K ف1u>_ Pym{A" #acc]@ il3 GFFЍ9b!xjHpek5aKk 4d>}ĶEST{C9Pcsh7zـ@FV&7 ®򃯾G4gl6z]gᬽgF2K]m/Mn]$ӧIE+PtףDۂibǶ9r]j|D2m p*(iՠ+fF&L$l{,ӴNs q!&i5PsG-^ y!k\3ܑ\Ft2I_$n_44iI#82.{\h0I9} GF cN´ʖ? /z9W˽.7y<P2H.‚59ʡn ;lx j(IJlg66C?A*)pb(YgPZ`=jpAñ5Q͸ojUPY{ׯ2|دB1 +-|z_n|d_]]Uv~2u2@3$ܔRL\jĿ"EۀW{fUAbm=bw#d^gua0~OqQy+Ԑ?zW_ 6({FIOfq1Y>nd];] ~_ +?l Mʞ"㍜3Q8PJJ0K"w %* RT@coXHdq.VT'#89$"^+/ T ŅFqezGZMZcoZbtCl]IZ'ߠ]/PDhِH4- ig.P!sD*R;?Y}~UQw|M^~LFNSd䤇%#/}*x ZDB='?]'/R>yj^ z{.,?^o8(cq:^BMMs;Ԧ]+Ymz 9KM[tjyenju`>~/YGS{m6oI:f p8Rn0-SO2ZCPM6o\6CiG5g꓇&\45c{/ّ_b9_sݻȇMԹ>łFJPocRZqݹ`U)tFmd|+rHv'o@$(v.b2MOI",tcpƶ32L}\y(UdZy.` v}[g$r-5N'y>;FYh3MU]$˥'#/x/J'(zѐ0v z"įChH.ʯJҡpП4sOH6S>)eϘ`؂xMs`PP ҆qDfHNL` ˭#, L$SGyZ$E4;t $(F㓔2˩ yBhq '`Q\M8i3.%P{.%RS[r5[r-d6T2ܟ2EzrpP M $€K+ɵ{Dp]Ԅfʽ4[/wQ*q#-*SfyF(`!(3K_KBNZF^OX֒s c^U^Wc!ְj){0pTԉ<|XuBr_r5(eD7c-H%NY#CM&5ƩL8%OĤ@aF08JiF!2LE4I +S9VF3ENhTRBM-cqtGr#\RXyUuxp(S(.@7 #: "&;D(efsz5J (=J)K>ZʧVg:]y][iRZ`}Il~(xe 5p∡#;?cXaCv`3s։9&ւn3+~|\@i;f7@LyN O`0O:0X;9*Eq# yw|6!Du_ .Ft6_,sޛ>BWE<.n-gaF]}{tWLpE+/-PN-0ѲxӜ#,q#9:N˩/PA@8BETmװ*wL7d|B溿,=V;P<n}1Cq!bY{v$ (wp-mRFuUUynfu.3D֗0rHҢɢ;I4hEBpJ[lt.K*a÷,΅}$tò;bӏ,N~ hlvri0TDA"C^!GVQNGp?z_u7f^Ri6r⃹ b:k˄7/Y ǟ=F0pLZ=i2"I*%)A!0$J|6VK +1%TRn0C"YHSq2 ըJUiUwUbi Zb+)㱉o6']Y]OZ8s6%4zGx4CNwt bv X$j''8`kH*Yd7+ z51NS" D)BȎ^aR({nin LSNzepzi% IF'iql9 D);+7SB;19Asrns%/,!8IDB3,￟LX?3@% %OQH',Hf4IL17ie1@f"5HJu""CfWTn+ -uTK>( \I5f<6!a!*I+2&#LUC*$< #0?7AX #wے}4iy}M6mn mf'r q</.?GKxnEǬ {%X9~uu\@7\W!HJx.}ײ^Ɂx$+~,_Tԕ^]ZJ;NĎ6aR'l HMۗۊa&ChJL4'JB =m%mdDH YA)W;E1ɢl#97^Z]"I)`Hltaycn+4iib^v f0q[Z[ޖ7ǚ$P&09$A2hƌ"#$2<2˘M$Q")Rѯ,׫b=ЁFrȚI0 au)%Uv T <%"mx-ݝ,?Oӧ]>8JY%c{Ec&Y4^82H(ӕi|@uyr% 3֜j.o80aʀr@^~%cR=>{]BhkD*~|}H-/qZwtr!!rl~Z#(̑o!9͊zpZ/oBf^RrD5MEmߋ_ܬf=?gQ8JH(*òeAM >A,a1΄dp6<2X' %Y#w0QL)}JЊ`43H Jڝ9/G"U⫔v_hy5JB1C @vZuLiwڋ|3{vgD}{Z\aperK@9k .J Sj .˻ [MöEرq%r-lo3ZrfxqP|鋿}3{B 6$imlXͅJwRеlP[nKNR/w/!M[ׄM"tN.qҬD9'bXr~FSn^t%Kwl"WZ[gҘKMQa@LJ܃ |&h ~[ 8UK(徛j(hBG6ѐMpni͓}I~%|R{7|Ɠ |{F(kxR?Ҙ'`6hOyi}p0k`~TT&F{6J?=ޏ7#2"cYJaV^s=vu7O6oWcqe_>y+CD*LRv\_e]X"MQ̞5FrDY0ih SҾzK Be8W`!ykes!_r,hLwK u0Whxpmt A\z6nP ] J6}\#A LH{:|^=OwRݓD%Xsu\tru]PIBn&KƯJN*w~Ե%i?٤f}@opĥ]Ul{0#Lh-@"]fŖ÷G 7Jڌ,vuk e?f<=k4 (q 孉v_4Mn.qٻFrcWIޫh`]o"}I`ta|[IKnv%٤*KIYUźB)ŦٝOq{MWB*%cVӽR1J`zL{1?Z-@AH ƄЁX!z)VZ\ζÊ z<..GK> =GqU~@\WWSH:@!NU\:POnbcq^*NlV.]~4j)Uuu"qJƭٚ[U{/^fC~x{5(VF:. $[M9EJ3;$ڈTzJ 043 C\Keg6*Uq,ƗuMfe־yWV[oQ`bLBjMr!\R2lsVf퍶RHUrlʌAnLo[zCo^~mlj mQoj)<ק7?o}zuF#j1|z<5R0+VKi{T`~-^v~{r+2ŧb,5^ʠ@SN.!sY{aoe uSSsd *{s4LIuZ{j*y!_(FHEi&A |`5N6E!i?E`Fk&5$F"&RU$-l<<ޤOOȇu}Vld*.>ll35V;[Ww* -,PJ~n4jQ?ըڐ@|ܹE05]/\ydRk.nf=5Ɋ)?{?9ׅ{]8ׅ{]uro}#!%4)Cxn ^/h$X2RibSO%俗 eM?t9Y7DZ1GM_\6xM>0PKͮ~s3nf[ؓ!3 0G WJ3II 2"Yc La%fl*JuպdbH,)Ms:H&Iǯ nz@K G"%ᘂ~F\L:@)4jE$_*Qm8IS^6"N(m"DAziSeթ<ρ:ZfE>jB$y%T冧)S4Z")U[*NXTjWzD!bJQ1zoxv҄Z(|`?HdY2~A^? BtG}Gjt`&>ReȎY6??P5OPsBצcdl[qM"Y"{sHEa< zxƅP2hUw-Es*Z=ݻ& N_|*v&Kvf̡̡Sm EDY
"މi"/P@ۅ7^]!U \DIU[m\3Z^#N&-݋dC4A}a}b]T]|"/7WWtI҅[/P Re{njgUƫNe(W-Ӯ O1&.mDҢ/V7 9&/b|o?' ~Y6_3'f8iՌ;@ď;ip&q"?3L{2m&wϻͬ-Iȫcjk[т!M:ls Bg7'eW=mλM r;ݥ:O{ @0pl>X8 (2>N+G+yWh݌_P5MRf!婳B\&X2̸J=u'# Sd2 d z,d~YT6h!qHLyfrYsjr!Q&,d=F ~g .]?=+}So<O @P+v/*,W|Sġ>_Ko3^Bg3߿bFǛ4)7!_wnuMܯǻ;|A. 4(G"|?Ss*g!O|RM4h:7pSA7F OQ.Mj8@%XejJ(p: WS 0}~#|~z>=04ޔm[&Io c/]p&K^K6tɠC%Wd.7u_cÒk{@sZ(0ic#2e3 @(m=Wɜ=e$ЎL^}ը_Zv—/KOC^l7OjC; 1 Z#p8UF8 q5Hbˡl/@_& VI1Ck| Rq$omET\3"eT"R)Zkl=`,?{Ls 36Hen=غ*q Leq{G̵u*d g#(Լzy k(wYқցzуJװa̛|yhrO/L)@Tg : 8}si)gy֋'nJ0B8p@4B8eRDS+8jqvM9 XfU3ne,'y)feDRg%ZYDFpqLzjP?Lt\˄)F,CcQNVQfD[ Lgf\&LD{*b2Z'2e:02%GZ>%Ҧƕ`^*u1Pt& ]jRj=je:Q,|Gj>["X[)n6XNH:4Tg?#f7q͈]%;xcd9XGn< Íإ16> ʉ+f- 2E^A+ ZqG^pMFKӞ,ka46ߋZ~ojVjY6|J8gмK9V7 J Q}M)$o΅^,rC?PHAҘ9-:[>VY$+U]7^]i=@|:d.(퍘o|ԫ-Ry{Z-3/hX?UJԻqęC$6KQ"!jR827 n>~ bHVg48jԂ!vވ`,f)Ăfh"D<7)=b%3ٜmHtAu~|TTGZ G1FH (;*'_@akONiQi'8TIʺ޵$M$E ;?>ᄍJ-pPSD+:MD8Jyj;Vv3}%DCʾsoK[2TVT" Q-hNa:A)k+dk<+'g"DcUbӨ$k  VJBT1j_d4$ǜL:͘}#`וg[Nj"Ưf(3/}%bt΂8 T}9/ }e% dYSG2e QekW|RXې!zIdxn%VxxT jJh_~<"IX-0Jً[e6ag5kЭp gZe5nɤb}QT:YJv082,A0 )'Ѡ\ɦjUq Cs9!F1"S* s+b" zK)1u9H"RU@ y4<) $DmB9(b za+"}tAd;ug05=xXI9 VpܞoeGaC<86#f =*:-gxfg_ fS8A7#B9&tF: $̺s=T !~v0HTØהZ"Tovh~ zTR5l֩mT*qL{Oh'i׀cpnUbGX%k"Q@j*G[SZr3_L|nV"BWl\dRG5L73ds/DٍI \JPׅ* !|&njipY@f!T+.V̀GQ뼪WrGs,XM~*Vrr5ߗmF{I]N>rA[^ډ+7'/_g/me5^fqȩRkܔƑLd y_L\І['錡f6#Rʍ&:'hcYqr69"'uhچLDtDmV qCcLGk\ޯwTiVEW&nǖFەg@ӄ8:`vO*H0!Feb똢Ɉ4Vo𔟫߯D!eAs5qj(z/kV7iZny}ȧe5h4J{WIt-{ǯD`@[^[J{-u-G;Pśږ0cR~U-6xWZa X64}l'ǔ_~̇-VOoI TlIzbD*%4\r a.2n.v7 q1hPsq{J'{N?CV-F򻱕к-A⧮#&uh{ͅ}+ZaaoIYWo&qS}0xhƣjϩiQQEK9* _&RF#HnL4=5SnQ.qQG% x9X}nz)Nr@\]odz:ty 5Ee>=ӁeA<.5^kt&wnaSo׾EB~]ŘJ [_i9͌RY.,?Nܴ3wm#5YViCN VVWkۑV%mЦHl)H)T@*s\wx:V'xsb>$s14ʉ! P Fu#Tp 4:f.x[uˊKΊgkWI?@8➂]]hCnPg** 'Βsu{j/HrDd;KD*X1V(+S "9D@X+RHͿ9U!(PϪ@=NrPAg ~;H(UWZ5vKqL71-dж̷PKR%D9Ȋ8 &+gJqin9ԏGh@|/2]H)2,ċFZP72_g;st5ȼQJ-0 +0HIz]M FQ1bP"BfZ-&=^̧GڜdٗY/gUEpFh3VL~'l|oJϠ(Ǜ2͡m~gGA(g>s!(Ad uW._`uZ-j>biP3ֶ/wԗ9ms9MޔMUT( jT<+{L9@]P UA_5ؼ-<&̾bzဗ͈vk3`Zz17)oyc4c%FalV% DQ!*cdQʑE\ C HDQp*.tEsv7o`7@ Q~LhB$ ]68(}`t׀AH,ߓqȖ:ӌԢdaS-)ܱyL#PZ\"# eڬVoB(ϑ䁕[C^R "bC|A_|򙲔Fep%N*1pU=ʭ^44G,K-4FHpK%Q!%00 UɑE}|6'bl/V4|9JCDe.}i:W5",3" Ȅ,xFF(EĆS[x2OZSR3r߾Df 9\hEd7 6p0Y( lH.13. =((uPL]X;QPHpΡ)#k.A,)Cc$a UB&^|D4ػ63Ϧ5HB{ BT"2uhQs(ǰE'T!$yE׊BRU+5)91 ƠRB!ZVy i*(63kDoDUA* k*"Ρ`J@Q2mL_XRz-jxQvjݍ?oVIȼ+a>=R `/Y_o7~x`~ŏ~jۻW 5 k?^b~:> ?#^i4_g[j3-g3l $On_1c@fp5_㑗߲zf T[M<q*TK9+Xa]P!U]!VD(Ka KQV ynƿ7uG$ jDBA!}hawT7ň%dlz7 ]$D~^mL"pow~92=؝޸xA[xh p.aAܰrָ[=a33y[ 2`DٺAzN2ڼQ#:-\kF7O'(d>, 2#'2  =DdtAB}{CN6졟:*&C9I 9LE( Ĝ[X%#7IY_Io:L-Rւ  4ns#w/- ^={ Icx=Ϗ(=B |ZрZ 1z o,SSZ,/&D~|nS0K~,c˯B~yV6\vq)H% 'wd%i+]=uhs-(QEZL*2IWؘG~fF .!qC =V+p8"84znΟkϚJ9V|YĜ0U eŲ%am\g `,q 2ve9$CP*O2"Lۭrpg\R~a*$a}zRc Yw#6Ϧ\PUe-ӿ VVA *WPK k*" X.Jap BCWJOQ2^WTZ)dQ bHXǕ, T*KBzA9ɫ@=^2EJp(.6lG4?(fx  X2,8y)BDbޚ -@`lnwj}=nZ֡d8lWX 3  ${v&9(9ЃD:N [, E/?맷"vPaҟ E\%@?~M+GMN^bq{IP 6(|NiH{xb "νXne%N+b"ALR-PKqk3'#xj祜$s̸$k,|\ir~D%Qʛ)j}|z}_X"S{WG^LOD%7GM\q9n p(wؗΗw/tQ8 '9I+@NvKEoLB\dKS\VYΘDQ9%v^{W|y0vM ;C Cα3$En:+Oj7g 'ТRڵr$JXAd>Ʌ;ÆAe;wM<hiDb2^Fٻ"$n `>oqA>opcD…WzƋ^@].JL@d< ) EĥRzvb9""̳CmQ FTL}iwsmơU10@93aL(~m)4d i]`BB-o,dL9H| BAˢ+P_J" n I*,+8kQS FDFF8*M-5 !LrvP,ǐc9ةzCc/BFr $HMՌM9<5N p-MNY6s|,}lQ@ƐVzҳUM[/lYK68"Ɯ/mK1jǠCV 2 އ?D)6)Bq^ ݗW^ va^ 1BZpaĈV9NG. iqRL5?-ɩ[{~{璳/@j0 Gc̅:H]g($HR>, ;uIIZ3lAD%${#׎{+ ޘ_j19B3s44'7X^Mؤ[Lfz策+RD׷IBQˊb;hYeu ^e{O>{jɃ=Zu;:Lh~_g|^^]_C@ _ .*>vj4Uy7}b?c/M$s}z묘?_] ݴyV赎3Wl^'!{>0TUR{z3rIAqț~SĖtгqBo փmRg \8e!IF>}kݗBBvsL=ku4|77&x8CPH6J{0[sMJ47#كTFOpjy)%#Q%]% [ġ AJ@Toэ e'DUEMk[r/-C2cI@ ݅q04k4^YdVꂃXy- , J jHQPֱH¤FZ˞vMu> 9CԎE͑_s/[!Sn-U^Y~֍]ҮUqU +kȺ=`ql4J\NS:]9r(2_hMuμ)dW&appU8%7;?Ӽy`RhW('*2DKW{;Al]|*/PMkɘ^ƫoAd9tdE {0VΪV/Xb_n3Tk:\󲖶dmqqOg~+>\Vmχ7|X};EyL|XהyCʃVْ=NB(dEDNsb"CNsg"H='ȄuB.& [IrE2:GBm) \qVRK]5C'e .LW Ah̭$An&?LcrRr PHP0IPKVD0N4Z*!%JR%XVzmw W"-c6~yN|BGS' swJogW-Uˮwڐൾ]0oC p}`78}`ONv1z(C {7ctak0 Ng% $sO7t}%(뷳 wPrJސj讓LiZBBN4Q!;3 p?\9w*`$?}EBa'HހASbCj_67wWhz?>zهOgOx3[4?5+Xeޭ7ravO2>Dc6%yug5d8)8@FcW^\4D=5*}XUWE2\·cJ$Y'G|͏$l5fvZSĔQȆ鐕&ʈUɦ1g<U$ҭ|-X=4|8l']PmsDk̄neսeA2]]^\uc1/ONv1vJƧ4[Н$N}q;pX [)m=w.om04C:1 _zH{x/la Btktkn3uM$d?ܝ@ mz[m! R(:9PrH(L$#k= e s (0̪!Fc(`( z 1˲aRk.O0 ,ilmlAܮ@7$ 4ZCOl&CAvEgV2B\1oyϽ@q9G=үţ+$iE;`;W׾T(XEBbĖ9f&j|/o^%|v# lkZ.n+pk2cqhWǏkł_ Fyjp]8֧Y;u{~yi f_?YX\_~qۡ5$'AESha2mI~{| /t4OFn*H2BͮA>\N̹tnXX_Ke8{uJP#k꯳n_} n\Ϸjqlr 0][ڷ&x]i,mQT jv M]fѕ[uF fd/-pǡ|:ߪ}_;>Ԝ*Midi 5s8;3UZ|76y! &7Ws%1X 1zoR짩E5&xhA&GM{݄ O&sͱP'&CZ"hIlr2q*P̔Bdld"4(>pftWdP'"咊t"/]%/ kB:s8g2ܟ}+Umп__|ɺdX&a+u6gn>ɔ_zS~M7mSz[ZUo%Q%k^!(ʹ,km(J,eJCgx}~n8Hjk*ӭ7MN٪#'SOXU{l쮫<%"ɨ9 !bPi6gBW\0 yîQ[DB^S>" 0Iy*Z ?(_fkB87@9t)Hd (lʇҳa)t3Ʊi!yBUmY)Xl]J7U) 2V%jIuMD~W"h%3+ƣ @`ye]׋gAxv2@~ nSK]Pդ$el:sk]*/ڀ*,0Dօr0UkdJJij*Qrf2I>B7fCDUآXTPp%kJge&wIHbs_mK>,6gQS*ee(D7 15`찣cp9A9/@8d;5y$tؠCBN+y<.|zp2WpFz:Յݻ_%@zk|b8@Fly gԽj@oִ'p Tx%. 2/x7#bypq;efN"buOQJͲvdN.A9.<+Ouicx$k`W#!z9\f.wpף^bCP{-vŚOA**tP+"׵cbaWժ{QQTUd&m\JjbVB=N,{ˈ.?dt,Ib*.*>jpSּ0OU)jݔp)d_8qʪbR ߍ3k 5  k*""Ro-IȫcrzΈ#Wޔ/gwAQJ{piGmbD )T=gPJɓS|adHFt44p]5:Y3'!/$d \%s@^1 ԘJ Fj/B -.X(.:q!;!RiRZEF(ntQjH *R [1oGo߻/>\=={jyo4^ $ӏ/O?>{yw?o_?uچ6Y٬~X<է3S7;7Oģпs맯pl7&g-M9Qe/}Bq;EnɮÇU@2%}+h2ߐ%ϛ[hGgoR{@"?qNQqgbX05޽M\R7mغI͇<|J.^J-6XɠCc%b+Ǩr֚ m0v&$Ե?^0-9#[]6b8cFh{φVH%?rÃ(Ւ'G_v)1te2wwG'|4y20$[7:=E;-V $ałbd&birFs5rQBBWJT4w~[۵8AHKXE;R zpցۀRX71g%kW9( BYa(ozb[|'WýoG/:9u|ĬMBq8R@c!ş`WxJ1GC-yzƀ"yT~+c@(*,18#LjsiaCO sB q!3qNv> sw;@p1K9ٛ_ҺmP_ay~X^e\MLaR0]IT6VsjiUu!*!֒ kʥ~(\;Q7}W!v*$(,>'D)pn:`wC.)SoRR+lX-*N$m 1xr؞z "KN8Au3}@$ H\nA96!/~:ͨ0ɳ3+Em 09 [CǜF,k]P'~G9˞k*|q#L}i: ~X8EluF0EtdXv&Y%ydw%?G2`[BNޝynpֺ<) h8BeaH=8}nDF%тNZ>9=-D0=7tA=Ȩ=&Ȱh!H;'g=$߄[ֺ\ @xxoo6Q0'D+G{oR}?Ύo FEЗޢ|"Y)q]d.n؊eKN6eό~k ؑ[U"Ybs&%I@ bĜy0ҖNjN :bIbЍz0P%[C$ZGqzv*J^*a$T3K.h#N6 {>goLx*Q`J#>eS]iC.W} <dk9x#DUE?j^WRҩ51њРɈWtQJ' p5'ghЬ]l+{ރ 5}Cpמx}6!*<IJƣ;~$S魟̇J/ ^Z>-٨.)8;gU7$*$juwtGrѪ͈z,鴌.U@3?e6 fF߭e64#kVh[h j#fhrl,f+@Z:&*OzxhZ<|&BB =C i#;8~9I>X?=:ױE:ǩuq -0PjkhĠVlmP㐞e1PZA0#w=/L{ɹe/y'4t/i;2(<樇Q fε qqL ۂ 21Ѝ|QЯ3 _OYRYP:kE96D P]$9kpչܓ]o^+?3֝6E5/B.Bz{qywkՑX_\wڭUxmz$ץګtO~Տbς1mTzn]r[}bWBh=zCv?}5{KO#cJy2b(^hk +HyC[F`WlMAP1NS%R``XnER?ܺme"RAO/<[|٠au\ җ˴#r#=Q {KYyW]Qm . }\oi4yF|uGK8X&_0k9"x㢤RT OԻ됏@=T)/QocGZVYyKlwQcNT i5:x0cC$oC3]|JYİ( R8Pr,ZD)Ip?X`:-trqͦW146peQ άliKW|cCNYjdrע,4a'A7d|*1yX~b yʇC !'|0"< @pwR#%Bg͏}mf|(~@$ Ξa!jN;C%mޔKKȅԧs]M>q9>g )YϭuM>^j&j+?gQE|_ߟݫz޽ݫz޽jλg)tE ( Ѧ uu֌ךJ"2ZVld,~k+o-T\ @Eɵ) Hkw }(/lǔOyC$9[hh 3 8YC9zD9v&vXD|4^ fhrF#p̐#W~Oyx>9ֱ-GyLȷ~zt0'B D 5c4c*Eڵ|-~'CL,1G0lc>tWV!Π6t(2ce|طΆׅ~m>$61ɐwS(c)N;ӈ zV`%j*XY+~qK>`D BUI|qDMaKhϠ$Lu A\@y!  \zv@`H9mBfFMˊLJ?oȺ}2R*5<0UjUeTAcBK[|hr>˄M(DEw.: HEcJVBZ>T,x r}zOREdc*t3EWaDX^0Ը*}HׂPJ6\uŘq`Zdt- *&fhr8b6FLNv[?IvK[l6nI&+-QRBH bѡk~16Jdfb47ACc̄,+(?|T?qPBuqg[,iM]] PhDcLb~O-a {2;w6ti3E.ӫ̲xuwb^6-I.97|`qoT3l,%ȹeȯatXM;.VK5:G8CABQ1& U9&%+[ݻAq;ޜv9'I% "ň3%1vvlCmb}U^^5JLRͯMI(|2T`;G-Ar* @Z{vNY(sމ!`|hε90?nql0V~^кOTL^R`v>!R9-AP0I|ڀ]ikRaJu%:WwQZAҁ$A[U}mp΅`d  :b#)[ rk0be@̹VCͅ*%@3TO8t(SRVRe#yJyU}SUT\f WKqVbǙ5*?O/z!Vb)KO]6[MWgw&Txty"\-7_ycWounSf;7|-`QuQ_~݋?ǩ &.OLP\w\HrZJ짡[Άui1R);AHhƫg>~9Ig^?=:S=sؼq#zɞ3ӡ^p y+{!1lRc[B㜌 h},z15Ϫ{̲j$O QY.IO ^)l*T`P7n7-1Y_* ,ϼH۩) B'CB9h Υ3PKuotPAZZWglƙU/<9jReWkp& /5yTLj,*Ux[ WJIh^^-)INKFH9s$x)›fzi' fI&$ܙf¥Ӄhr>Oc9 N'瞢QYCv)?&& c5_q;J;_v B~db!8(.L;njbJ "H,NyX1y|sw[{]Z>D:ʹ[W]9FA\%Asb=ͅQFApmT]=R 2A!J{Y8>UhsfQkʻQh(‚bL~T_߼wdqi-MB>0 o| pˡ*^-#`Wgv f뻳Ŧ೟J)qC84c/ab,7tw HLIOWc\U/)j "\ǙdV,+WRKo" ahxUVޖ93M;*B} dݱZ.6$'iýKIt]+὎zh8p Ǧwh?Rr=_,@nYlm}}5=?jQO^_. أ{Yh6p3t̩4cr,PŔ;{߼"J2C,~%FFZ+,K'rS O $Ujf-s9DH' M &m(9`0VTnT·42 aoDi F'A!|$}j&`0BCTJ=*y)'k,b KB:Iׄ6į̆2?nauH? _#2UG/3o?tZ$?&W;Zd*X[gVߍ׮Pdti12lgtwDy\,]MAϓ #&qAv 見P@kU:BOzavv?|&jߥbrI$j}'I"M @9I=AYmMA +-|`!Яᗿmid+ګ|˧~2:B FZyiƎq&sKI+\Ł1A>+CAޘ#}eCXUl%ycp ?5(N;9Pjq\+ k<]P.^-jjPj %e{R[Hc )ݡt)0a476v/Ԯ/VfPm(mC ;؊3W5*jM$#% jk:`hS)tOǩ 1FγƔJ1xSjp!/~ %3@Pj'u Dj%!ޡh89^sMGA90;}J=g&)W1l^5Nma1(ndg9mO81B}7EmAk0khD2[Rr(B$k︲ӎ=t=Rv(sr?|o?>QZޞj(96 (9ʑ+D$Zy-A4j,Dg՞O߆.ɨ!~POvz8ΐpW3k_uX!@9]L$7biRZ1EɀbKT&fjg~%ri>Jދԁ.G?V#dp/Ysç"t@ C,>Ka9e:N'>4:AjzJQ X[O vV! eo`IN ko!骧~2:(-V\|ÿ&HG2ې[} X,3(nS t#)FE1j?{9߸.<[9:X,џ6"Oͮ%MV>w8=ߞiG iWft`35Px'`e,^Jw(-3ZR3T t78RDxajtWBxU5g  sh ( xU-Ea򟱝8P8#HI5Oz}0h Jjaޗ*ˈzs&Y_Jcs?~OMZeEF/ 3}y9*.Q€׽\' ޻X r%H&w`$pW93`gqItQ97C\ z_Rz8aОK޳P 5~h^7~I2P'ٞIh_g(ajE9[hi路89K!Z FpF!Li91QB+}9)#"\1!+TB:GTic9)*I16"Odʌ&pj ^v>#uoދ?X((c;)!_[湆jeY!ʢ2A,.6( V8ØsIl!U"X2עs)EK$hnW; gD4VPrk1M- L,R=!-RA4 ]5I#  ) Ł8B+PH)S uV!9@gG;(iW_1mnL$ecBӴhB&iC}F&i+)h}8 (JҮTJ*vډ"4+k)GO5Rռ9w57Usuw, Ǝ>ϑi#fֺl]2Y,ϳ˸ٹ{>?\NUgy~_Mf:L")AQhz& Д'^i# (@3hUl=Xb"B hhp%%RrU}L0b2~Zňys&OUK8Dՠ-uc޽/U8֪DݖifT!\fא,y&Hˑ ߖY=d`ya JUn=^h%bKsBrQR95ђpBܡ#&P- j[OD_! 0S**͔.<`)7ȏx,gh[Əmc7OctlÇc7C2㷑c;7/\9F猝klƇ .&ïHtOܹdf  Ft<ɜS&>UGkyO?]:xC6ۮD#Yga,l{^,$yMDSD@;'cCƃe0cȕ>VsߏN_N$㏴zT ɧ&՜ԺA|V`U0XFȧziOv1_Vz>,R]P?Ϸ_ej~Nn$KRp_^eSt2?]ݜ|`'8w,5#̀"S W@'Dc"~\3]?P*3foח\lѢG)bn<eZƍaf讘U:s29tOLgS :O~S.kN'pu=/UUG\QZ8GL\Th[|0x|[N9 < V{dBWg‹söUA`ÃH+]x*\|zp)'j0u蕮5mqe|KLP|N,nT!?1j9bv[1RY CZYAHhF+c w5T$eIӨV'zR1 j(L'ìE{A]EU*^(k)q,ք7!TSR+$>bYG*BrcĔuq[@:.4䃫hNi~{u$[*BT'uvDKS`-/кu!\EStJQvuc=fA"Ljj88^ȃ2YtІ4䃫hCA9Lad>$7ܐ7([a_/\8ɺvq\˅daH+i (aJ4n#Y֯! g?λ7[CLtI=vw#VZFDV)6Өn̔ez>E\d .ܯ6hZ׷[?qo u)2c5SXکʱUϋ+][o[G+&;!~`g#1%*$1߷iJnPil.YժOyKG}ZIaO,=jֱmnK1먭ңfi]&B0v~J0Mܧ a;L!+y6 re;k&Ft@QG jYY裶^_KLХh:ѕYYxyD.q먭ybQLK"~:j$szbQ.zMĮ3>IKG!2a2Q8⩘)'íy*w@;X Ā6%s*Ed_^(!d5͸8 "2()rY^# }ϴN.3D#3e4{ǵșѺ@aBa#&(eƜN *auԶm/AX_no'{ַ.y_y-ƃ`c/QKXoa\׍R~1ԜTR-q퉥Ru,9`Ʈy&>O٘zq>3dE" 4*]6 BFsJ^yVrG8*Iz_9 ݛ7Tq i*v׵+bЭ(0Y%;GY:'A06WZl¬UNtu}QI"4t)ebؠ<VPf+e.lPITs\N=~fdr%#u|7<{'>WiUx2ISJNJ1 ^bޅ *i?y鑲e:(܅ *imJ:JZQ%{l~<֊@A+Ji:}8fK2nכ r.>5.آ?>ەBPaȻoۮd^"5g$rp&eTJ9__¼Ԉ%MS'}ٌ"i ]}-n.|˳]¢Bo&1 ǯS?'Oa_h6Dϱ$eP_FoMÎ˳|?ߔi4*j|j n=+ֿv 'WQ'#Q%鴟KUkPؾY]`6=UۇD~up2ax"}UG(cNR[nYW59ZP7@R<٩eߚ#}^ҴSG'I'Zk@>Kxتեb ֓,Ǩ63)R4!eUqLn5"4XΑ|"PfkTK^%/W, C(|2\qIxekI< 0L\Ja.U( -rۮD0ՇcʓD*+%_oy?„&Ѝ>Sv8|LoJ8 _h` APj&r9S(Eo˾de4яD [k4:KsRj@ Q~F#MDjJk%WLdɱPl,)F;yF.b$9h#=|z_6?ڑ2B͎"^/|]Kjӿ~m5pXw_7ɂ4\O&/9魟W$MIaUATџ޻?#2Ζy;oJoLH5w` P 1s쿖s2La $dz㥰__S[c۳V46&,Lf]wS`57}LT:8 T( 9ɧozesEKSR{+O_[ -/=oɍf}z>(,Ŷ35ݞBO`dXx,>QryU9WUVB665G]M+G_3}6=#|1ɰ?>ܼۦ+'jJ l%xfrVŅY_&fL\,ۢ'ъ/E7]_]l͓w~P}{}ߢ= Cy;7[&o^o޽WI?giΟnRX(~,%[9/Z7?׿Ճz=>hRX槟]a<;WU?6?]D AZ#r[|MM5"Ǜ7!h˳n±uγ{O*-: Y|-+$_AԺdhv.5dsp=Ѥst#uٽ/Zcm7vG 5ZYX6:2Zy\VuқirlH~AILBX{ n~wMhYA{ҔEbvOORoo@~to^gU&gxL7{#[^1M"i9T Y ޶=_MʞTdZhxhL^w1IJ>i^O Z{mew\>Dx/UhaeػJo]ΣSB9\|5-Ay VbX9UY&H[J:a-EWwmjIaԩ 4(0lk%a20 $j:{1PHʸ,|FB%} B WnH"(kUߐKE?iO*/D5-<]t,Q0QOx6F y=0{ VSq &J!@-(^&C=c|de+;Aܱ`aFe1&H<eRuKlAK$:.kc 29e CHfgO)$A9.l5$Ӟ6rעI2Ŝ5ZI@RL&9M8^faUPRisVdi1(Lcv=лgo9ȀR2Mѱ`d|4fl5A]Q?l hXJ>ER>JAb g H}H\I,JͷBLȼIqT'5,&y?{mj^٘gpLռ3fS&"p\$K:S ](xĕ* D6n4nh9/s0)9At`kA5Cyo]?.S_ecBˣ^qI$!q/S(KVk&*x+գކw)!?RTÆvK8ZV>}WAhcΐ Glcղ@G+FƜbNKv<֣0w [$G]!#64G#+33 pCb"e-ssG56CrT2IdQ.U&iۡS3e?$H$3H01v:"-%{S t=lQ"hG^S~  5g9)XfT(9ƒ~[x qA]wfY*H\nu!Yn4U$(4VJPhIJI q9J8*G*UPd7.bf,qƪI8#x`+QS8cdIbzve .cS8g tr\ە%H)X ,NJ.wT`:DQ|7a2aWQȔ z-SOxߩ᭭:R>QZ# W ( 4CXQBWkޱ ߄Ӌᇼ61k[ު~{lc]vb<;aҵ;!Ox)tҙB:ߐeض?UL0JD%a}];1,I(fx; l&G8ْ#H cD58HR"wm 1j`1rt@. k 4e o( _3g֌Īt@i$/pA &S@*LYveHRВI PuDSܹh$:2|B؂1vr .N`rǷ2Kd5;\2 ӛKfA\2; Օ,$6= o=ztf^&#=p̹w7ՕQ]`bKqcﺃA"D^گ:WLh[(W 5Q?¶v?oN}Io_k t@a} w HtH4KDB(0%d:!pjmm}9oׇHGvka ~0u˜ܺq53? 2{9cKS_>) װ{_s'Ucg-hĆ48QnafhHQF=~،׮YǼ-Jqq!Nh0ưQ[R 'bo0䒐?Z-LYRR P((\*M*%ͱpC%6PvDA(Я Z-4DjѧTAT#p AO[6I6lGR5E7:r r80캥)Œkd!OhI&2yK(-Oh<u\~{v ( @8ȠHed&9"`ft"cZz"B)+QK 4!ʩF7pniWytvq.3_L ZY|žgX B?H֝0G`R;aj}G @s߀ Ы֬]!d^\@(Hv_ݯRx2Z6؜ Dzu'=c@5AI5A~2O: pѢ >чecRfIy.sl=R ['Dh 5}*rDJҰfQRB@Q*Up@<-_8 2,r bEq}8@t!*(iS/p f߰Z<\ ? nJ la# W%.nTcVîuU:sr6ZM-FmV$-3}RpY(]S[HL߯w;f 9tŵ? 6!'ߍmҫ:R~Ho茼L5QIQBn1jэDM^pn rG`YHF8ʁd$t,tBj-tu-wUTwykO1xJ(0;4XA"oJvi^NjNሚ?kٍR80JПjˠP`A!2)8 )iHWD^.oW4swS$FiH_ؑa$<pFKc+C0Ќ2A9\'Di#Dn)#jkP ,AISQpAZaـ`\RQN$QƁ1L$4)X~9֤H]2%pLA\U6Z~V*N TidB%jU \! eJ3f(EbGQ",O$-K ,!(R cABJZ~By Ղc<&nT<]E+F_:Qm:{m-=|&ᮡhS_ hLy]ǝi7='nN;hS/0vhvK!!q)9jhUbQF*2B$ +΁X,2{^Kl r)Ure h᜻0D[%3!wdwzҽߟ ^֟g?;QYVV,}leUI/_'E7Q8;u뭼-BѷYw0wPYl6]ܘ3H|y##?s(zu׈Όzoo'GW5?Lk'/^什} i͐q;MR!7P ts$.xQ dD2v4 RXLTm1X3Pk)b,5sŠх `.ʾZ ( !1fSc%kwKEvyT.A+"f6/xY7 W/ }?.jA~{#+lvYec@NW ?4#2O[O;A@ωBI=0e9g.VG?ۥgɨK-"V7${߮>HzE/ڽhj٪,z(&J+R&b^2Kjuř\[ ϐI/5{kVZ?ʔYL̨_ qG*yګ6G7zbjEv;f 9'A)>-eEC/\AM>sXgܖ|kMߛ)&K3`؉nvJÌ] ]ԣGge2UQqԤy*ҷ{υmĕ;+-BeB d&5%f g!Ohiԓ6' O6mߘcgu<ó†X]rFatXk8q]A$y)QV۳z'zG7SX|ZO5[ONعO`U0@=e6wkAI[sBV?w!w gxQB_Jm:\ HJ/EsuX%wx >z!4cт|7CIK,ӈ:.`1_kwÐ>-ҧѣg 30f]02MѶO_~ř·?~Сm-3 ko6n `*x+2-BG,ݘ*a|f3iIW;NÒ$Un< ߐ)ʴA뤣O6? nKwO=:ZL߶vvLax'LOrŋd+|>^`֣G_ɘ#6n׸85 #] Fke"#.MyҾ;P˔"̦OggUyAw=l:~m\%:] ڗk(Mf!HΈ]͜k q) P>@DA-pB+dD"V&q$p`N8>@?ʘ 9%]ψXx,bacl;C~猁rfgA _> !8}pn8_!zc?9cxT3?{WǍ MFzpX܍׎q(pS"lloV7jv7U@hbDf"^x0j"z;nUy!!\;lr@r 5ܑ~ .Oc1wPahD|A`yoI5:(83:$jr#YN.\`sx`sYqeJQ3bUߨF?o}Չjf47I6ƶQzptm` "؈ͲSOw.cXt ˭[kMb]EއY[\9+(N~= Y|stf7_~G^Zwv~}'G?[M;B9gy;;%C^P2h 9V{ՒKٮ[;} eDrc;ŏPR:BG%kJ[#Z&h b3"eC* %#Bڣ]}6jk6-֣֨1d C,DQ6Fm'0C1ICUDA2*xBc'F?!Ri \۠N+Gv 9$Fc)/ <ÌuYT;L:ӃjxH3d1Lgcdwpe)PY,ea΁&F' 0 m#O{T; 68hP"(Z:a_ T` yJ$IhUN&ϑ8y\3ܪ1eP"yk"q)pS*_%eP@P>*(?ש*Ш%xt+X5dԼKh1yO&!&p7^ $?*Ԫ_ALPD5HHiP(@W5H& PZ쉔r._~svDrIJh/DùeVnF V0f-[kŚ\Nx?ŻU5nWzi G73nw uvgSu^ov64hm&V!WL1+FnU._KZʻlz-~u$ "2O/ċh_-_A2)GKEN*t>-&4W֐/\Dd:_DBҤ9$*Zh iEr*\UjtSn- cv_'#/2"{$:4M Q%)*dQcXt@!yU /-u3}jjR rI78^?ʺ_?9~ 0X4ܜ,53^73h, Q5n,~z|Ks7LXTvj,:vrƢF4bF@~0>Qm` BU56(8MX;P6($ZL/,5(8]/ > TEu2NgO-U匳,M)UuJ:A`ZFA/{O5ڗ=]jF0u'yJdPHv4&x 5BD* X%hzAPh4#"<_{R4:,D Ƌ8.9%frzjIaih~Kڄy:[ Vn2jH.E2e毓[.sBVA~v BK-&4WՐ/\D˔"VoWwYbLSg^NM$S[o!Poc缡 ;D5%QsPx#L|^1sްx$N46ҷ^=Ҿ򷳴\WEl龩-Ld Vխۛ2%w<7ZFt4^?tc/6w /Fsh)j}|5\pOo^i}x?,^6T`@ǁ0Kb8+Qy2Z@``C+JW y!%o15FZZXq0x[+ڭK{bdYbZ(Xq᚞zy%yKLZ/L0=^'%Re/X? ;X"q/-XkUՏR`\$/wl! GU\H˄\ŐBR^KpW%FMqʫEѺuqsTX9mpWo]9;?8 ѻ+?svW3Q;'Š P?-L"+AVsڜGsIo^GR)(QbAKv\3߱`74@ToF;9(88L acz kVzBP=io>\uŞohYo oԈ7t͊Мs z hJ ycŪTqγ Pg&YjY h=chL$EyЌFL hF^Jdf$3 ZՓ vV\[3R/XV}snMs5&4<9%IX3Ixt…G?JEOx%}*CS;\Kl,b4Kv4?I _{QipY JT V2l> Ֆ31:M.ЭQLúf,L$}'f$D/ &bPAۈ)I󨙽')D%Np˝PPp:-X'6""RZ߄pBEHxNY#UA`?޾%AT`SfcZ?ќ+,/ZvwuOW~ǬPmoD %>ѹ^M`yɿ"nOJ$`M~M{WdM6[-y0^*V?IbjMoSZ*{9ΐ1}, h\-oEC³ET!ہcy3_KsgtAYOreyG`C]@ޒgvƘ %qJ*V5ܺs<ڑ$hkUAHF;e2thTYxW: +=+_X 8ܸzics'3WMǵbLl$l 1hh-wۦ\0XP;V(VFhhƀ9F@ë@TI~*X8`ܫ~C:ū aX#ԐC"~ٖ"392L)ZwArGJ}i䪨~j ˪m؛ !A?)W@7k`O%Q=ULׂ=$C3N3dpC+c2YMNy]frͤǷ5Hp]>aP[ݜXK%$ Ub~59OaPL6RrY6z3G2C6W?g*HDZe1}RJG+uyJ𳋫qAמTI%nTb:H:(B˻B-* F6ty9 zXJqO%~(y?iQǐOYg @ҍXc ΍A; 1=8)+C$"o&ZsO˯~dnU7ZJ~x?MHd_ߜKt.;GLegAY˃{P\V,dV%eVB}j1ӏd;} ZXK}E+Z\7V X'Kz~\ĒV#n,m(.)X^ Kw@`uWdHW(-Kmư(`IB)ݶ` vB|gPL3ƢTRyr #SvhU+eudP2Ix3nNΔ Xe?P=ynA}m65Yݵ Rw[=h~RUYcp}s~yq}LTβ@Xc}0dml]˺o~[n2enCJՈ|8lVL7 $ygԓ-!{dTTzib%)0!~Eeqĉ՝+5`_h/[1=zYLТ׻k_R.,Þ9"OQ]sD3_l@Gn 2qmm+ݰKI$c Sٺ볎`JWVv~釺\|uhF^pX;cP2ֹQ-b`o9bvWb$/Wob̆dv鳿 $6Kr! H)$fԖ(} Ыx?տ\'=Ig.%29yy[!; VJ)}Gv;ށ(5ڭTև|"Z$SNn$v+A蔾#GuڭTև|"Y%je]"^1 ?XyFcCZ%4\z4hMZ8󵪕׍nHڅQ֛ACδDq,Hg<R{%C )1/S * * ZwSX:zUu_(e[ߺeEjQYe:&dIu^e7_~ f%˯J. ~v  tlX㻛e!,/r)x8mUhe5_|\ۿ+t hzf":] tJkEvHL4!ˍFujmk\TlŜ2$Ob`۱w˞6Rs˴-{*O32dmʉ%}M͑=4 `k$@̑=EJaD(-{ݵT2lw3a MQ>64e=C@UbNeHDWcbpg㟯Ohe&~ǧ,A`zMncTFa%%]%U6 K%/f} VS;&SA]ua:z$X g4_DŽF_9t}=U&?~FB ^X2~ /BT7qoޏ?xX2cgsu߮Cc_MlQ!,ʙ]]-'_>71mBKT:LaY\d)̈́Fk) Ni-7ۂ-µS:HXq;oI$ j#j2y   _4S'4hlXD:z4' rnpM]2E=Z$ Pfq1|Q7ecn5yQ#Gp1JZHD*㔱hZV6vFYеiUcA2z$y`L鋊Ɣ4 Z0%ASLDJ>[t"aӞjlu:AyV '@vyY7?ߕq.wxۡ[؞.qK7U͌<鄹/ү_+ )9wegbgKz1yk3$S `m6-0eim+Hj p9yۡ7ZN_⚲}2r"AVʩX. m'~H[%,DtiiN)QЖ)C+JȬQRSSYN(dv+Gr u*9xGtgv£ٝtTXBobn&)+\V2eiIޛ8I2 9XT~Y>'Lrc.*ңifj Lzx|4-Ku|wOʯ+*Qp{TwPՖ!\Gk$' 6m@# tADn{^%N k|*Yť N-*+x\ok[y4`錫onsb O 0 &ѡ0W; ^UX|XC:H@Fg6nǥYza58mh %FgD@[OגYH۾[:P 8͘a8_bXvddJ{Œ%Z0!X-TG ycfKG͹\: @oc=rB,pXF-jLlNFݹfZ%"Ȱ-/ 0:6K#JU4!) J!NP#3qt:Б(o"YMbBE:ځ3U $JIW0` ֲ}{Jc'olAm[PqF) R#gnj`Ji":n󻈧#nňj>$3"BsݐĆ!k:-Xu { Yb2il5쾎~B?tQ+S1ycϦ"wm=$X)4sjљy9PV^hmCw4D@U2X-➙ “7[z<4zV+}S=ot+dQ'CZ Yk̝&~[|M/ΈQLW[ RQ·0nb{%͝jbDBXYxlLH#WhM,Bz(`Bm8ZP"'$s8ZD(i=jV-ɖNR,,w:5Z3Y4WH D%ALywPOQ.%P܁i@e SW֡Z[Dm 6JΣ"`NrC{m)0fsABͩ<_ ی%TD8Ɠ1bXD83e37-E5,"s(yc  2@&ǽKkiX̆ FY=]_oDZ22Y b;f8yXlIH,ITUiCB C^l(~LD`)m@W62XiX:<900+Ie>$ M H EyJXk4,; <k!8@ w\0yܤE)H`Sd^38 N r h٦gv{ 2H~iJY1R- ErR* 9nW$%J!IWW&KTRB;hF䛴6?TkzC#hz#,igQ8h"廞w>Mmb3e^C͜9au>xdDcou}KV}'J8k;H*+le|_nx6 ΋9o3w؅6,Ul*{-WH!dS8XD+rfC'6g Z--k*_1o28Ó(SPޤ py-4᪏j _QiN Ը mr ]9 ]I8qe16ċc8#.Ϗ 7K% 1F_ ȔG;! ^!wq:9KhT|2ªtBY+ׅJRAQX ggS?"U~Ÿ). q ԕSE2:QIAyugP%pn^,ZU۬"\+v_\X\B:T{L --זћ4:FB $̮TLt"!M/} Y= GK0Jt>!Aa$ש P GP Kz(jEoɝc᮲M+Q/gU맍dĶsuKlX9v)hQ`BԗSZ$˷DfqpH,/FHj܎Wx6k~?R_җR/F2rՊ5q<,ml{*hWv JuY%))tD$ ̾Jk{FY$ڜ]Y?iIM.VW2n4MV7XfMBbS:|CG͆m2W6\b K )=%v8[9P]|<) swBM/?FnΈ5w49 )x#e9w 8@ G0FGLܯ^p {;V㵓>HD+Rxruy6VA9qBk0 -:]RZULԦU]Ǜ;+ݥiM Nۑ<ޖlWPg򭠕d2ocklq]bx|ℒ%Fc"J) r6\%=UwM lDžY,Ɵ=~9$ JUxb2-+HS#U~36/%eJV f^V6 5B?W2Ӣ3ͺoO5A]۲" /;|9)OI,\まXvAl_Zm-lugۀ*rFTܾQ%ƚP GMuB֔*B>5޿YcT-SJvǼcUZZ8GT+ZZZ\;?7=ZGT#ҝR TN#R tJ5jPӗ̗ -tJb_ZT&nԠAq6A|`+JT,_[%=wLA&UDH?O GWjc]1PuY`*4\')-2ק՟ ]ZnPsO5dh˿pP+pMc5}[,{?#_Xz9݁8GTLicjj)eQ?!Gpϣƃ{3w3hݚ4}:%㬰˳+q<.fvy*=wH9'q?G?:*xw$ZlJ?>w&,Vk6˫ˑȟ/oE8tJ1'彛{Le.F_.{+Rټ%|5F?jsNÌ2r~uzvy |ߛ}\#%&G]+hxLhW@M>͹݅8qcBLӍcJ( }9{8K?:n+ucP0X-&[t2j Se?.'W`48:^5;gճZ{w/5vb_a'N9[yžܟJPTiUVǨ>ZZ1ױZ*M=-f֍TzZ:gǵY=-i|舖"SZҝRޱ:Z:}QT1.?|x7Y?7V}'yte!,W|V)"'-#:86R߄N"; L: c( ۝ұʫXc|˖n6Yy=\ċ{oIύgrbDVEbtkCzS* ʹ1s2}#m(gB,Ul^HI9^06ݔF>Eꦴ茛h&ꔦwRCX!c_<*^")`3"i'C `K/gd!^$@9V[Ua D0ڡUVbac4/i͋)X%v :pFYm;U vˇ#ޔ3+_>莉չk? g2e4:jD3#7k#ǭ)|:->I/ngvo|kϳB0D",}{;ɬ౪j f_a&%rq,2cgLgԚ=JYL̺Hr=k[,sZ|o@J>$SoZgrIFzBlܬOOUdb҂oIL{MU)+lIFk51R ߆˫MX##JޭmY6h-qVWd55A+Ԥ#K\*cNAp`q!)p>Jt9a/aUfth{5y+sgiӞoOB<۝b! A٭&]d1/J-[n^Y^|j~[^F5\n+/ʷ蟸䌑|=6{̞~vI&Jjf-э؃b\|ŷ6{UIZТh|#YCe2֜ ["I֘ca^ pR|iVlUMnig)&D>n,ߡ& z=|ªCds>:a&[Pr7]4Qai0}0Sid8rt`%OQsQ'U!p)oN$AI/#KT˚GPr:e ےmM{mۂ&L΁FPOKT?{WƑB_3M"v`# ZWMJjRTniDVWW'vn~hYvw T<+vhJdSR`Sajq\v+$Jl^*AX)%yVJOJ_fR\Q0To8/1 !(-Bq S }rjf--·j.9֫-򬔫JL>GeǝmBY}4h녺_eiyR?,AN]hGbY9no'}?!\ϛMrRufrwm8-!*A! "ْjOWZ|]˭qR&Cûܽ, yU%m.x ~~P&GӈkGh&[R8zC&P v,T%Lje^wZPc9)ui-76X*G4(,VAJU%DBBԊ!zxV3o XÙ@EB"n F$x)Q6J*BWb!+`Es8s e&AT` PmȕRU۪`8f2H6P? zAݻ~ n8?n'NbZ)aqa SJb`y?,S-<̼zFdy8LJSajIԱRjVW[a fB !:hߓG42% 5\\4$V ^8f19iq2@qjC__ݞҲ-RK έ/`QE,/IDWMyݔH!,[$<6:uc|ʓ=-R#(n1lp擵q3kW.+Jev:2E%tT"P&| U3E9ĝgnS~L ;L+fg%sׇO*k~H|u$ZYuWqz????;B9P/*(LeJR,Q&n~6sϳl﷼ 9/:를ǁm2%X ?&;{X{iW|Z<:~XF8|b*!Y ?M1Ha̼PIT(AM(3/-51'Ab@¯=0DS4?Ke=Kv/!?柭dԫmh%FME2'@_{?rQB)ЦzG8\ګGC,j/7q+&tGi4X9$%Qwe8|륱gnsa`lXp8'w's;.QON7^M.tь|:ƃWӉg>p 29ȁ2B_a#8zwBЫڤM_w,&VJq>FS\]Ǣ b IsY2;a05f<F͛kf5^WE}Ҩ +Pg(޹WIFM:0[ח _js-.{U*1к*NF&g Ryiޝ[h_xɄ$1'(_oB/.vywd7e ゜ys9szXYeV#Mq9I5+W'> #P]ϋ<d*V:H&JiM)g F<@y|9!-G7=..M``h0TA'0~X*|  5W6ꆫyl7czE}󅝌.}kR-!®Kߚj o+sZAw-ɗaLV(M^Ihme”>+KL85g$k)f1$(G::"^*tp\MmߣmXPjy)}Ys>*rNJSuK (p MeR-T~M&J|a1ɋ +8О@!J+cIXv:5HvyKTH™ҍ4]lRhK%)Wk Nmњ)  uvy7*x-4tuT谄q+biף!T9 揬d)1*+5fms7p V N{$ J*v!0 !FEuG]jm؜jZo#Bi h @ ʄ+9!Gp291@8wQ%P&@r'U$STwwReJ _Fт;m 9M`9vNںFp}D,zpzF&S5BQ} Q z傈ֲ֓x/$-+9~Lj˯eՌur+I_o[EdQ\t}+.]n˭!~{no7/5nK@Q- E%qT)*dvKS>H͙>x~&&SpKE߼-Rq::{ ,$m(q(aD1dĂ \e  tUOm׸2[=pa>*Js `PQ+iY1ZXD4 [nئi9V#n p~}6lLfI(ۜPD #+wHoDT0Y!*Y@ #r2+<ǟ(BN ѽ ca3ۑ#oER& *o_5b++dhhzyg<+c KESji ѴS= V Q0 ˨`f k 18Pѥ7m7#p(eʖ+uw E` Dfhm8Gc_Li2|Ֆ%ZfSޘjfMe~m+c &pAo\!$K8$.=*nuލab.RY{{|d83rr1 NO U}L+Og "4$׉[T&H(q[Cq[6|yFȺT ^?VU59trw'=@Vue=ω)WfNvE] pm`12b--COFpF/kЦ)ui-76X*G4(,hN.REcUҒC/An&hC1Hjpd:0%m0Hpư# Mm·!-#΂n>S.U/| 4tNʋ]>X׼EVfvȏ 1qz71"?FLeyvaZMw4tL44k*BT砐; &$.C˚vBC2WBA ͫT7aP2N~K8k%ueш5ؘ֪jj}7Qxh՞j^KtlÆdU&k4釴-2*!=oWaV]%]z A%\R0>hǺ0gpXxr8hBښ7թablLGiӿώΎZ3zfbyT9BqmeSKaD7FAD21h9␵<<3ua!߸mJ,J+BRt-ΘEU[W z$߿oTJaeky yb?,K0F<NsXjQNacU$b,L,tFD֊&x a+zx|y~s]3 OU@us衎_vA sw϶ډ_}1+PC@W@Z.J YXfx,F)b0JtÄ-f޵5qK ?dL5֍$v4I֞JroR^߇nNle`4,t@j"B#"tiލqnt(qEwUIkmkКxa1JYr!kH&a2!Cv1dg`/o3; ۖ|ֵ(Z_Gq$VkhFwdzE]~#1\o-6}[4,-}p5`AڝH0ҎklŇZ@P܍l$#X?bWҔ`H~e/*n]YT 5u-K_?E]]yY:G l=.W;z𖏿/sݝd_kMǟC0(Wk߳W}<w7\gf?@~|#PٛGy9gNvah7?݀o߆wuwֻ3ϟ5C~[xϾ4LH!USftfȬQsF4[@W7Jx5nρ 'WhS$J:Mh2-$ ?=qC7Sݐ>F=*4L)t xɹ:FGX>GscّL; d; KCbAjuw"CM8qKΘOé1`G&8&7ۉKMB:R>R1d^XMDgvyQ/綜:6?_HO/n]^4oЋU/^4?CT:' H6AA~#%}nH,{DF)<䑻hOl>J B,O3Jel^/NJf}FGVxwef6ާHEdg1c)U`0ٕ rWw֫m\,#iINa- ֕"mKM#557Q<)(`3ԧ zb-Q=b 9Kr--`,:rTW)$ 8u qEuDw:\k, ʗW"x_554U%ʷb,J6j$Xb6G^&B0B?WѶEk+lа/) cemںh*nku8 NPaႤGH ^v J9fP@7L?<<.ϱDyaN~X}H?|;xè@GL`paKM>@Mvﳡ6bG?|OIԓړ./)2};=~i^)m㝯kv1ZGXy_Q"u;nսՈ/*{R `Fw6-~33- mǯEŘVXR 6$xJ$%(q9J.MiuVE4y-D\ b\')mŢ'ش7c#]tO1nb6جbpJ?76F29ج$o%}y,^)>CϨƵ;;gaΘǰI@%aMQYz[2}IFQZ(S=l.adczp6c;mX?j(Hꦺj-\|ePE"qt^cGbϔOUP$tVQ&zc4UYI. z sS kc X%TȢ1WPG[zQGz#.yyCg^H#a,a}F jpIbo %TuDBRQk$ϝiarκDq1Yx@ɦF|[Ɖ4MZ-MQڒݕg),zp0t%Q]Ȇ؇$@9 C*0XwQ]+Bʊfd*e)Tz Z5o3'j$sEZ0[q^8܇\$nu6bɕYBjS q B[!&UZB[k x3zf:A3j+;:dcV"1ٓ͜?&) nH^$i;ЄaЊU#f5#`ю %N`f{xϿTroLga_<${yZ4'Lr!$fL0/ 8v'9~@ .[#%MFd{)l^0F6u&VnPnA(/=7:KYt^(JP#N9׿k~mGy7ǝ\+&xZ Oh@FfKlcV7U[`ޞajc \ackBB<[`\jwPBlivNǻ|4[Ӌ[&Y4X(}rNaIaU|80Й|ˍY$G$+qgsh/^z+* R ejJ?5jН.hg 9yj8jށvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004037144015133646520017705 0ustar rootrootJan 20 08:49:24 crc systemd[1]: Starting Kubernetes Kubelet... Jan 20 08:49:24 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 20 08:49:25 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 20 08:49:25 crc kubenswrapper[4967]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.524977 4967 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528174 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528193 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528197 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528202 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528206 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528210 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528215 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528219 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528224 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528229 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528234 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528239 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528243 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528248 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528252 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528257 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528262 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528267 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528271 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528276 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528292 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528297 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528302 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528306 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528310 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528314 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528319 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528323 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528329 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528333 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528337 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528341 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528346 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528350 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528354 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528358 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528362 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528367 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528372 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528378 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528382 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528386 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528390 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528395 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528400 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528406 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528411 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528416 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528421 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528426 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528430 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528433 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528437 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528440 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528444 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528447 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528452 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528456 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528460 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528463 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528467 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528470 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528473 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528477 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528480 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528484 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528487 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528490 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528495 4967 feature_gate.go:330] unrecognized feature gate: Example Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528498 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.528502 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528861 4967 flags.go:64] FLAG: --address="0.0.0.0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528874 4967 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528881 4967 flags.go:64] FLAG: --anonymous-auth="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528887 4967 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528892 4967 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528897 4967 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528903 4967 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528908 4967 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528913 4967 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528917 4967 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528921 4967 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528926 4967 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528930 4967 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528934 4967 flags.go:64] FLAG: --cgroup-root="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528938 4967 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528942 4967 flags.go:64] FLAG: --client-ca-file="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528946 4967 flags.go:64] FLAG: --cloud-config="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528950 4967 flags.go:64] FLAG: --cloud-provider="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528954 4967 flags.go:64] FLAG: --cluster-dns="[]" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528962 4967 flags.go:64] FLAG: --cluster-domain="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528966 4967 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528971 4967 flags.go:64] FLAG: --config-dir="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528975 4967 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528979 4967 flags.go:64] FLAG: --container-log-max-files="5" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528985 4967 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528989 4967 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528993 4967 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.528997 4967 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529001 4967 flags.go:64] FLAG: --contention-profiling="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529005 4967 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529009 4967 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529014 4967 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529018 4967 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529023 4967 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529027 4967 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529032 4967 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529036 4967 flags.go:64] FLAG: --enable-load-reader="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529040 4967 flags.go:64] FLAG: --enable-server="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529044 4967 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529049 4967 flags.go:64] FLAG: --event-burst="100" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529053 4967 flags.go:64] FLAG: --event-qps="50" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529057 4967 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529061 4967 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529066 4967 flags.go:64] FLAG: --eviction-hard="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529072 4967 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529076 4967 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529080 4967 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529084 4967 flags.go:64] FLAG: --eviction-soft="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529088 4967 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529093 4967 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529097 4967 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529101 4967 flags.go:64] FLAG: --experimental-mounter-path="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529106 4967 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529110 4967 flags.go:64] FLAG: --fail-swap-on="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529114 4967 flags.go:64] FLAG: --feature-gates="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529119 4967 flags.go:64] FLAG: --file-check-frequency="20s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529124 4967 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529129 4967 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529133 4967 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529137 4967 flags.go:64] FLAG: --healthz-port="10248" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529141 4967 flags.go:64] FLAG: --help="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529146 4967 flags.go:64] FLAG: --hostname-override="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529150 4967 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529154 4967 flags.go:64] FLAG: --http-check-frequency="20s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529158 4967 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529162 4967 flags.go:64] FLAG: --image-credential-provider-config="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529166 4967 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529170 4967 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529174 4967 flags.go:64] FLAG: --image-service-endpoint="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529184 4967 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529188 4967 flags.go:64] FLAG: --kube-api-burst="100" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529192 4967 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529196 4967 flags.go:64] FLAG: --kube-api-qps="50" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529200 4967 flags.go:64] FLAG: --kube-reserved="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529205 4967 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529210 4967 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529214 4967 flags.go:64] FLAG: --kubelet-cgroups="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529218 4967 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529222 4967 flags.go:64] FLAG: --lock-file="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529226 4967 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529231 4967 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529235 4967 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529243 4967 flags.go:64] FLAG: --log-json-split-stream="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529247 4967 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529251 4967 flags.go:64] FLAG: --log-text-split-stream="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529256 4967 flags.go:64] FLAG: --logging-format="text" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529260 4967 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529264 4967 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529269 4967 flags.go:64] FLAG: --manifest-url="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529272 4967 flags.go:64] FLAG: --manifest-url-header="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529278 4967 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529282 4967 flags.go:64] FLAG: --max-open-files="1000000" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529287 4967 flags.go:64] FLAG: --max-pods="110" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529293 4967 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529297 4967 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529301 4967 flags.go:64] FLAG: --memory-manager-policy="None" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529304 4967 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529309 4967 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529313 4967 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529317 4967 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529327 4967 flags.go:64] FLAG: --node-status-max-images="50" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529331 4967 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529335 4967 flags.go:64] FLAG: --oom-score-adj="-999" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529339 4967 flags.go:64] FLAG: --pod-cidr="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529343 4967 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529349 4967 flags.go:64] FLAG: --pod-manifest-path="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529353 4967 flags.go:64] FLAG: --pod-max-pids="-1" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529358 4967 flags.go:64] FLAG: --pods-per-core="0" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529362 4967 flags.go:64] FLAG: --port="10250" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529367 4967 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529371 4967 flags.go:64] FLAG: --provider-id="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529375 4967 flags.go:64] FLAG: --qos-reserved="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529379 4967 flags.go:64] FLAG: --read-only-port="10255" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529383 4967 flags.go:64] FLAG: --register-node="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529387 4967 flags.go:64] FLAG: --register-schedulable="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529391 4967 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529398 4967 flags.go:64] FLAG: --registry-burst="10" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529402 4967 flags.go:64] FLAG: --registry-qps="5" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529406 4967 flags.go:64] FLAG: --reserved-cpus="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529411 4967 flags.go:64] FLAG: --reserved-memory="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529416 4967 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529421 4967 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529425 4967 flags.go:64] FLAG: --rotate-certificates="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529429 4967 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529433 4967 flags.go:64] FLAG: --runonce="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529437 4967 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529442 4967 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529446 4967 flags.go:64] FLAG: --seccomp-default="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529450 4967 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529455 4967 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529459 4967 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529463 4967 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529468 4967 flags.go:64] FLAG: --storage-driver-password="root" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529472 4967 flags.go:64] FLAG: --storage-driver-secure="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529477 4967 flags.go:64] FLAG: --storage-driver-table="stats" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529481 4967 flags.go:64] FLAG: --storage-driver-user="root" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529485 4967 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529490 4967 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529495 4967 flags.go:64] FLAG: --system-cgroups="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529499 4967 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529506 4967 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529511 4967 flags.go:64] FLAG: --tls-cert-file="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529516 4967 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529524 4967 flags.go:64] FLAG: --tls-min-version="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529529 4967 flags.go:64] FLAG: --tls-private-key-file="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529535 4967 flags.go:64] FLAG: --topology-manager-policy="none" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529540 4967 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529545 4967 flags.go:64] FLAG: --topology-manager-scope="container" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529550 4967 flags.go:64] FLAG: --v="2" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529558 4967 flags.go:64] FLAG: --version="false" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529565 4967 flags.go:64] FLAG: --vmodule="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529571 4967 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.529577 4967 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529736 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529743 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529747 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529751 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529755 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529759 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529763 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529767 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529771 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529774 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529778 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529781 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529786 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529789 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529793 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529797 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529802 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529807 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529811 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529815 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529819 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529823 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529826 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529830 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529833 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529837 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529841 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529845 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529848 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529852 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529856 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529859 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529863 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529867 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529870 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529874 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529877 4967 feature_gate.go:330] unrecognized feature gate: Example Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529881 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529884 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529888 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529891 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529895 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529898 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529901 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529905 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529908 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529912 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529915 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529920 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529924 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529928 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529931 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529935 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529938 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529943 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529947 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529950 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529955 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529959 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529963 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529967 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529971 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529975 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529980 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529984 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529987 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529991 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.529995 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.530000 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.530004 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.530007 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.530139 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.539102 4967 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.539139 4967 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539244 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539263 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539268 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539276 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539282 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539290 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539297 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539302 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539307 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539312 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539317 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539322 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539327 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539333 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539338 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539342 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539347 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539351 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539356 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539361 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539365 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539370 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539375 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539379 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539384 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539388 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539393 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539397 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539403 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539410 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539415 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539420 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539426 4967 feature_gate.go:330] unrecognized feature gate: Example Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539432 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539438 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539442 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539447 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539451 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539455 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539460 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539464 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539468 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539472 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539476 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539480 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539483 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539487 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539491 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539494 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539497 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539501 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539505 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539508 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539512 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539516 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539521 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539525 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539530 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539534 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539538 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539542 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539545 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539549 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539552 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539556 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539559 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539563 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539566 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539570 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539573 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539576 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.539583 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539724 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539730 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539734 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539739 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539743 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539747 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539751 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539756 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539761 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539764 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539768 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539772 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539776 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539780 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539784 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539788 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539793 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539797 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539802 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539806 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539809 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539813 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539816 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539820 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539824 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539827 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539831 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539835 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539840 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539844 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539847 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539851 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539855 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539859 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539863 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539866 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539870 4967 feature_gate.go:330] unrecognized feature gate: Example Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539874 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539877 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539881 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539885 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539889 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539892 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539896 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539900 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539905 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539909 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539913 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539917 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539921 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539925 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539929 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539933 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539936 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539939 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539943 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539946 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539950 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539953 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539957 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539960 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539964 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539968 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539971 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539975 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539978 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539982 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539985 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539989 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539993 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.539997 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.540003 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.540305 4967 server.go:940] "Client rotation is on, will bootstrap in background" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.546194 4967 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.546347 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.547244 4967 server.go:997] "Starting client certificate rotation" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.547302 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.547570 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-04 20:55:04.27384557 +0000 UTC Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.547745 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.553423 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.555475 4967 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.556820 4967 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.568797 4967 log.go:25] "Validated CRI v1 runtime API" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.588052 4967 log.go:25] "Validated CRI v1 image API" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.590478 4967 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.593338 4967 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-20-08-45-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.593384 4967 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.620891 4967 manager.go:217] Machine: {Timestamp:2026-01-20 08:49:25.619420093 +0000 UTC m=+0.204720340 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:127fa9ec-49e3-48ce-8fe1-c4315b50082f BootID:4107ff2b-f75c-4198-af98-8ba98c029f9d Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:87:71:ce Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:87:71:ce Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ed:42:b6 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:6c:1e:ad Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:7f:b7:a6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ec:30:bd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9e:3e:f6:24:09:01 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:19:f7:b6:b2:0b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.621179 4967 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.621406 4967 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.621968 4967 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.622444 4967 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.622483 4967 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.622753 4967 topology_manager.go:138] "Creating topology manager with none policy" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.622767 4967 container_manager_linux.go:303] "Creating device plugin manager" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.623004 4967 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.623060 4967 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.623442 4967 state_mem.go:36] "Initialized new in-memory state store" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.623852 4967 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.624540 4967 kubelet.go:418] "Attempting to sync node with API server" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.624569 4967 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.624616 4967 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.624658 4967 kubelet.go:324] "Adding apiserver pod source" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.624679 4967 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.626479 4967 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.626830 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.627094 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.627300 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.627148 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.627392 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.627517 4967 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628024 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628056 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628065 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628075 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628089 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628098 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628106 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628119 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628128 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628136 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628146 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628154 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628378 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.628857 4967 server.go:1280] "Started kubelet" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.629291 4967 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.629439 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.629810 4967 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.629996 4967 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 20 08:49:25 crc systemd[1]: Started Kubernetes Kubelet. Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.631728 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.631765 4967 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.631889 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 22:02:37.626803566 +0000 UTC Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.632238 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.632315 4967 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.632366 4967 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.632526 4967 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.634454 4967 factory.go:55] Registering systemd factory Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.634503 4967 factory.go:221] Registration of the systemd container factory successfully Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.634834 4967 factory.go:153] Registering CRI-O factory Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.634776 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.634856 4967 factory.go:221] Registration of the crio container factory successfully Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.635115 4967 server.go:460] "Adding debug handlers to kubelet server" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.635122 4967 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.635306 4967 factory.go:103] Registering Raw factory Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.635340 4967 manager.go:1196] Started watching for new ooms in manager Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.635978 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.636527 4967 manager.go:319] Starting recovery of all containers Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.639418 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="200ms" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.639755 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188c643cb9dcc169 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-20 08:49:25.628830057 +0000 UTC m=+0.214130264,LastTimestamp:2026-01-20 08:49:25.628830057 +0000 UTC m=+0.214130264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654245 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654298 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654311 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654323 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654336 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654348 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654360 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654371 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654385 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654396 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654408 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654421 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654431 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654445 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654458 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654469 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654481 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654491 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654505 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654516 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654527 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654538 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654548 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654559 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654569 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654581 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654597 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654609 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654653 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654668 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654680 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654722 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654843 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654856 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654897 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654909 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654922 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654934 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654945 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654958 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654969 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654982 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.654995 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655008 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655019 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655134 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655146 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655159 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655172 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655183 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655195 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655207 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655222 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655236 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655249 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655261 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655272 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655284 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655294 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655305 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655317 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655328 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655339 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655349 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655360 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655371 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655382 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655394 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655405 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655415 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655426 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655437 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655448 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655711 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655727 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655740 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655753 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655764 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655776 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655788 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655801 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655812 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655823 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655834 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655845 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655857 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655868 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655881 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655893 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655905 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655917 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655930 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655942 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655954 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655964 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655976 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.655989 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656000 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656012 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656023 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656036 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656048 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656059 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656071 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656088 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656103 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656116 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656129 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656141 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656154 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656166 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656179 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656193 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656205 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656219 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656231 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656251 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656262 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656273 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656285 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656296 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656308 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.656320 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659174 4967 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659235 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659268 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659298 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659324 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659357 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659386 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659414 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659441 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659468 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659494 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659519 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659544 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659576 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659602 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659684 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659716 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659743 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659792 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659822 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659851 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659879 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659905 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659930 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659956 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.659984 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660012 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660038 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660080 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660106 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660136 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660164 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660192 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660222 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660248 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660274 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660298 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660329 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660354 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660380 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660406 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660435 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660462 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660489 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660515 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660545 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660572 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660599 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660669 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660698 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660727 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660753 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660783 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660812 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660840 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660867 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.660994 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661019 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661042 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661066 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661093 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661117 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661141 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661169 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661193 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661219 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661245 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661272 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661297 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661325 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661351 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661377 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661417 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661443 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661468 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661495 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661521 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661547 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661571 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661596 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661679 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661711 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661739 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661769 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661794 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661821 4967 reconstruct.go:97] "Volume reconstruction finished" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.661839 4967 reconciler.go:26] "Reconciler: start to sync state" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.665589 4967 manager.go:324] Recovery completed Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.677053 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.679187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.679232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.679244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.682525 4967 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.682556 4967 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.682586 4967 state_mem.go:36] "Initialized new in-memory state store" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.688864 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.691830 4967 policy_none.go:49] "None policy: Start" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.692828 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.692898 4967 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.692958 4967 kubelet.go:2335] "Starting kubelet main sync loop" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.693074 4967 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.694000 4967 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.694057 4967 state_mem.go:35] "Initializing new in-memory state store" Jan 20 08:49:25 crc kubenswrapper[4967]: W0120 08:49:25.694151 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.694239 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.732325 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.739751 4967 manager.go:334] "Starting Device Plugin manager" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.739796 4967 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.739808 4967 server.go:79] "Starting device plugin registration server" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.740174 4967 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.740190 4967 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.740405 4967 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.740490 4967 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.740498 4967 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.748496 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.794226 4967 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.794312 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795509 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795734 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.795770 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.796847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797011 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797064 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797279 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.797881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.798673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.798696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.798707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.798831 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799267 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799323 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799890 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.799996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800012 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800045 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.800886 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.801878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.840578 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.840937 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="400ms" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.841701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.841729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.841739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.841760 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 20 08:49:25 crc kubenswrapper[4967]: E0120 08:49:25.842048 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863434 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863480 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863509 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863537 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863564 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863593 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863653 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863682 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863785 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863842 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863874 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863899 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.863983 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965070 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965502 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965766 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966218 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966434 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966595 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966122 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965677 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966333 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965904 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.965333 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.966873 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.967413 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.967671 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.967892 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968027 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.967831 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968268 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968100 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968586 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968723 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968794 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968888 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.968971 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.969021 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.969045 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.969156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:25 crc kubenswrapper[4967]: I0120 08:49:25.969516 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.042692 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.044022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.044066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.044078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.044107 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.044765 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.123133 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.132351 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.148097 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.151850 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d3a54bdcd1f3583a10e68e9f831f88787c4fbde1024ea109bc7ff74d8f4a7cd1 WatchSource:0}: Error finding container d3a54bdcd1f3583a10e68e9f831f88787c4fbde1024ea109bc7ff74d8f4a7cd1: Status 404 returned error can't find the container with id d3a54bdcd1f3583a10e68e9f831f88787c4fbde1024ea109bc7ff74d8f4a7cd1 Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.156650 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.162794 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.165304 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7f764fb13aff3394e5a932de7e17a1f7ce4934ff080fffdb753b08ef75424308 WatchSource:0}: Error finding container 7f764fb13aff3394e5a932de7e17a1f7ce4934ff080fffdb753b08ef75424308: Status 404 returned error can't find the container with id 7f764fb13aff3394e5a932de7e17a1f7ce4934ff080fffdb753b08ef75424308 Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.166491 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-215549d04d8333ca887fb002682785d1ff5ab60880968ed5b7b317da6f373457 WatchSource:0}: Error finding container 215549d04d8333ca887fb002682785d1ff5ab60880968ed5b7b317da6f373457: Status 404 returned error can't find the container with id 215549d04d8333ca887fb002682785d1ff5ab60880968ed5b7b317da6f373457 Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.175586 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6bab5c6e93bcf275385a2fd6ff9dd279eea4d79af71660c3f70a4f7aab546a61 WatchSource:0}: Error finding container 6bab5c6e93bcf275385a2fd6ff9dd279eea4d79af71660c3f70a4f7aab546a61: Status 404 returned error can't find the container with id 6bab5c6e93bcf275385a2fd6ff9dd279eea4d79af71660c3f70a4f7aab546a61 Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.177392 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5d21be73b75d0b6f52ef4419a0e3fb8d09f14d484fecc0f0f91bedf6f5fbed91 WatchSource:0}: Error finding container 5d21be73b75d0b6f52ef4419a0e3fb8d09f14d484fecc0f0f91bedf6f5fbed91: Status 404 returned error can't find the container with id 5d21be73b75d0b6f52ef4419a0e3fb8d09f14d484fecc0f0f91bedf6f5fbed91 Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.242679 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="800ms" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.445823 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.447271 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.447293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.447304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.447328 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.447799 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.631161 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.632267 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 13:16:19.8639278 +0000 UTC Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.673433 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.673548 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.699606 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1" exitCode=0 Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.699727 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.699843 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3a54bdcd1f3583a10e68e9f831f88787c4fbde1024ea109bc7ff74d8f4a7cd1"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.699941 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.701248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.701279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.701290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.702780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.702826 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6bab5c6e93bcf275385a2fd6ff9dd279eea4d79af71660c3f70a4f7aab546a61"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.705237 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892" exitCode=0 Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.705291 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.705336 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5d21be73b75d0b6f52ef4419a0e3fb8d09f14d484fecc0f0f91bedf6f5fbed91"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.705435 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.706588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.706727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.706757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.708038 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd" exitCode=0 Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.708098 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.708123 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"215549d04d8333ca887fb002682785d1ff5ab60880968ed5b7b317da6f373457"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.708247 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.709254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.709284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.709298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.710415 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.710574 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c" exitCode=0 Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.710603 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.710645 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7f764fb13aff3394e5a932de7e17a1f7ce4934ff080fffdb753b08ef75424308"} Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.710714 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:26 crc kubenswrapper[4967]: I0120 08:49:26.711989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.811517 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.811652 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.818419 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.818541 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:26 crc kubenswrapper[4967]: W0120 08:49:26.886723 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Jan 20 08:49:26 crc kubenswrapper[4967]: E0120 08:49:26.886821 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Jan 20 08:49:27 crc kubenswrapper[4967]: E0120 08:49:27.044016 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="1.6s" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.248244 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.249815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.249865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.249875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.249897 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 20 08:49:27 crc kubenswrapper[4967]: E0120 08:49:27.250375 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.633198 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 15:28:43.496909872 +0000 UTC Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.716909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.717184 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.717201 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.717270 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.720046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.720082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.720093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.721920 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.721947 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.721954 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.721960 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.723789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.723810 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.723817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.725541 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.725576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.725586 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.725594 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.727342 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c" exitCode=0 Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.727377 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.727446 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.728096 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.728142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.728154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.730211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5"} Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.730293 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.731083 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.731126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.731138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:27 crc kubenswrapper[4967]: I0120 08:49:27.747161 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.634083 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 09:26:56.552888854 +0000 UTC Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.661325 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.737016 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4"} Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.737084 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.738043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.738101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.738119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740500 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189" exitCode=0 Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189"} Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740584 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740635 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740652 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.740776 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.741911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.742057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.742090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.742105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.851377 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.852685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.852731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.852743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:28 crc kubenswrapper[4967]: I0120 08:49:28.852770 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.634713 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:04:32.962339696 +0000 UTC Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749102 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749356 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"efae4da5f2d2a14a06c78e20cc23d2811dbecf7a66eb86ee731a74e2c88ff4f8"} Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749403 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4eb232f2ebcd144a969f5fc3cc5b192a4454434917f0a7700d9f1373cece0253"} Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749418 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5879f9d52ca4ec745e9adc618ca74f9c1ecd395bcfc85b01adc06e5ae93a6da7"} Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4cc6b7e0c06a081516d203051e26fe1cb4a5febf9d273aaea117b284fe423c69"} Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.749930 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.750320 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.750377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.750470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.750481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.751587 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.751715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:29 crc kubenswrapper[4967]: I0120 08:49:29.751743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.583909 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.635659 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 07:51:52.199776988 +0000 UTC Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.754802 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.754837 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"09e324d3f818853cf74667a9c020ce195feb13f3535d6c29973c681e3553a3fe"} Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.754978 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.755697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.755748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.755765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.756123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.756185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:30 crc kubenswrapper[4967]: I0120 08:49:30.756211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.224729 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.636351 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 08:07:30.374501927 +0000 UTC Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.757877 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.757879 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.759861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.840226 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.840389 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.841673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.841710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.841723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.851699 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:31 crc kubenswrapper[4967]: I0120 08:49:31.917219 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.637477 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 00:22:43.630716094 +0000 UTC Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.760240 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.760285 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:32 crc kubenswrapper[4967]: I0120 08:49:32.761799 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.637976 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 21:50:38.314835342 +0000 UTC Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.663411 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.663702 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.665097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.665162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.665182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.747968 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.748227 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.749856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.749942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.749968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.762693 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.763763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.763806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:33 crc kubenswrapper[4967]: I0120 08:49:33.763823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:34 crc kubenswrapper[4967]: I0120 08:49:34.638766 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 00:37:43.13494772 +0000 UTC Jan 20 08:49:35 crc kubenswrapper[4967]: I0120 08:49:35.639147 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 20:08:05.345850038 +0000 UTC Jan 20 08:49:35 crc kubenswrapper[4967]: E0120 08:49:35.748873 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.436010 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.436157 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.437354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.437403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.437413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.441298 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.639602 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 10:23:21.058361614 +0000 UTC Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.769599 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.771028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.771081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:36 crc kubenswrapper[4967]: I0120 08:49:36.771094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:37 crc kubenswrapper[4967]: I0120 08:49:37.631746 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 20 08:49:37 crc kubenswrapper[4967]: I0120 08:49:37.640146 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 03:05:54.699921337 +0000 UTC Jan 20 08:49:37 crc kubenswrapper[4967]: E0120 08:49:37.748564 4967 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 20 08:49:37 crc kubenswrapper[4967]: I0120 08:49:37.983042 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 20 08:49:37 crc kubenswrapper[4967]: I0120 08:49:37.983167 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 20 08:49:38 crc kubenswrapper[4967]: I0120 08:49:38.230364 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 20 08:49:38 crc kubenswrapper[4967]: I0120 08:49:38.230548 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 20 08:49:38 crc kubenswrapper[4967]: I0120 08:49:38.640700 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 11:14:04.612008069 +0000 UTC Jan 20 08:49:38 crc kubenswrapper[4967]: I0120 08:49:38.672280 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]log ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]etcd ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/generic-apiserver-start-informers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/priority-and-fairness-filter ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-apiextensions-informers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-apiextensions-controllers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/crd-informer-synced ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-system-namespaces-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 20 08:49:38 crc kubenswrapper[4967]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/bootstrap-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/start-kube-aggregator-informers ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-registration-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-discovery-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]autoregister-completion ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-openapi-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 20 08:49:38 crc kubenswrapper[4967]: livez check failed Jan 20 08:49:38 crc kubenswrapper[4967]: I0120 08:49:38.672376 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:49:39 crc kubenswrapper[4967]: I0120 08:49:39.436507 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 20 08:49:39 crc kubenswrapper[4967]: I0120 08:49:39.436581 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 20 08:49:39 crc kubenswrapper[4967]: I0120 08:49:39.641495 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 16:46:16.859970585 +0000 UTC Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.543523 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.543762 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.545083 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.545129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.545143 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.568140 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.642354 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 12:39:48.604853167 +0000 UTC Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.780579 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.781277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.781307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.781333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:40 crc kubenswrapper[4967]: I0120 08:49:40.795415 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.643429 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 18:02:50.657899927 +0000 UTC Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.783304 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.784160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.784191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.784203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.981020 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 20 08:49:41 crc kubenswrapper[4967]: I0120 08:49:41.994178 4967 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 20 08:49:42 crc kubenswrapper[4967]: I0120 08:49:42.012486 4967 csr.go:261] certificate signing request csr-vn7b8 is approved, waiting to be issued Jan 20 08:49:42 crc kubenswrapper[4967]: I0120 08:49:42.022317 4967 csr.go:257] certificate signing request csr-vn7b8 is issued Jan 20 08:49:42 crc kubenswrapper[4967]: I0120 08:49:42.644581 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 11:50:17.285878514 +0000 UTC Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.023295 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-20 08:44:42 +0000 UTC, rotation deadline is 2026-12-10 11:44:23.224731614 +0000 UTC Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.023341 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7778h54m40.201394332s for next certificate rotation Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.217451 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.426090 4967 trace.go:236] Trace[927127976]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Jan-2026 08:49:29.020) (total time: 14405ms): Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[927127976]: ---"Objects listed" error: 14405ms (08:49:43.426) Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[927127976]: [14.405090969s] [14.405090969s] END Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.426120 4967 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.427151 4967 trace.go:236] Trace[1428850711]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Jan-2026 08:49:29.328) (total time: 14098ms): Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[1428850711]: ---"Objects listed" error: 14098ms (08:49:43.426) Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[1428850711]: [14.098482102s] [14.098482102s] END Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.427188 4967 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.429072 4967 trace.go:236] Trace[106532951]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Jan-2026 08:49:29.420) (total time: 14008ms): Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[106532951]: ---"Objects listed" error: 14008ms (08:49:43.429) Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[106532951]: [14.008995771s] [14.008995771s] END Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.429098 4967 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.431469 4967 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.434534 4967 trace.go:236] Trace[1855986890]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Jan-2026 08:49:28.641) (total time: 14793ms): Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[1855986890]: ---"Objects listed" error: 14793ms (08:49:43.434) Jan 20 08:49:43 crc kubenswrapper[4967]: Trace[1855986890]: [14.793121412s] [14.793121412s] END Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.434556 4967 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.456221 4967 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.456359 4967 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.457467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.457494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.457506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.457525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.457539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.475582 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.484060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.484090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.484098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.484115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.484124 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.493500 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.498026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.498048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.498057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.498071 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.498080 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.509383 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.516281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.516320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.516330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.516351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.516362 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.529563 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.532627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.532663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.532673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.532692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.532702 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.541304 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.541415 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.542684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.542718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.542730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.542746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.542757 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.633194 4967 apiserver.go:52] "Watching apiserver" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635122 4967 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635336 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-nxlff","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635654 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635667 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.635722 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635786 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.635831 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.635995 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.636075 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.636088 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.636142 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.636179 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.637428 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.637529 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.637709 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.637968 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638361 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638453 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638667 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638766 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638872 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.638969 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.639639 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.639748 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.644852 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 06:09:03.043232797 +0000 UTC Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.645528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.645665 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.645685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.645703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.645714 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.660674 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.669139 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.671758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.674374 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.681572 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.683701 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.685601 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52592->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.685664 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52592->192.168.126.11:17697: read: connection reset by peer" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.698227 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.709366 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.720306 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.732762 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.736165 4967 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748169 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.748679 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.762561 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.780304 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.788406 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.789947 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4" exitCode=255 Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.789988 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4"} Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.790513 4967 scope.go:117] "RemoveContainer" containerID="0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.792812 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.812311 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.822446 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832660 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832704 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832738 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832766 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832790 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832810 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832834 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832858 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832912 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832955 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.832979 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833008 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833035 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833036 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833057 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833080 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833081 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833093 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833104 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833127 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833152 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833236 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833260 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833285 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833290 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833308 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833330 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833378 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833378 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833421 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833445 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833466 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833487 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833498 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.833513 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:49:44.333492638 +0000 UTC m=+18.918792845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833535 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833545 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833560 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833584 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833606 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833646 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833669 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833704 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833743 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833736 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833781 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833804 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833826 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833832 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833850 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833874 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833897 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833920 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833943 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833966 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833989 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833031 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834012 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834686 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834715 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834766 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834794 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834822 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834847 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834871 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834893 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834918 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834942 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834972 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834997 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835020 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835043 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835063 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835085 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835107 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835127 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835146 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835168 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835188 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835212 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835234 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835254 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835275 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835296 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835319 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835342 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835365 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835393 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835414 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835436 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835460 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835490 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835512 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835535 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835558 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835581 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835623 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833894 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835648 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833916 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.833937 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834007 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834011 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834045 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834076 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834198 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834200 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834208 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834227 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834353 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.834389 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835955 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836225 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836267 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836283 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836296 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836408 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.835675 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836503 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836521 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836537 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836555 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836572 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836587 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836603 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836644 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836663 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836678 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836695 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836369 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836726 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836743 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836751 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836759 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836795 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836956 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836959 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837112 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837245 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837266 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837425 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837432 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.836830 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837602 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837667 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837692 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837694 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837716 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837743 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837766 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837797 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837822 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837846 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837868 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837891 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837915 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837938 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837961 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.837985 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838006 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838028 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838014 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838052 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838077 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838101 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838124 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838149 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838171 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838196 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838220 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838249 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838272 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838295 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838316 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838338 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838374 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838420 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838535 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838392 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838734 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838754 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838770 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838787 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838803 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838820 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838835 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838850 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838865 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838880 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838884 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838897 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838916 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838943 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838959 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.838974 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839022 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839040 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839056 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839071 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839095 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839130 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839142 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839146 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839214 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839236 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839308 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839336 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839327 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839373 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839391 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839407 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839424 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839439 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839458 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839475 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839519 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839534 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839551 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839567 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839583 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839605 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839645 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839660 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839675 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839692 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839708 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839726 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839762 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839767 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839777 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839795 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839812 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839827 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839843 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839860 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839875 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839892 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839923 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839941 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839960 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839976 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.839995 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22fng\" (UniqueName: \"kubernetes.io/projected/73537611-ae3a-4699-a02f-da4f5192d702-kube-api-access-22fng\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840014 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840046 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840081 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840112 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840130 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840148 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/73537611-ae3a-4699-a02f-da4f5192d702-hosts-file\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840166 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840183 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840197 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840213 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840250 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840261 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840271 4967 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840280 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840291 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840300 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840311 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840320 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840330 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840339 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840350 4967 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840359 4967 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840369 4967 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840382 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840395 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840406 4967 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840418 4967 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840430 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840443 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840457 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840467 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840476 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840485 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840494 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840503 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840516 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840529 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840541 4967 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840553 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840563 4967 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840572 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840581 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840590 4967 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840601 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840635 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840648 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840659 4967 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840669 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840677 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840688 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840702 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840713 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840726 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840765 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840781 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840794 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840806 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840819 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840832 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.840845 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841017 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841085 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841204 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841371 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841375 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.841917 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842038 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842059 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842083 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842093 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842855 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842890 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.842906 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843033 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843083 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843399 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843433 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843594 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.843993 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.844224 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.844364 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.844844 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.844963 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845051 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845329 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845463 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845551 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845668 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.845845 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.846395 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.846776 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.846862 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.846894 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.847002 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.847067 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.847081 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.847801 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.847177 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848154 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848170 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848183 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848270 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848555 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848853 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848861 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.848939 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.849379 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.849484 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.849694 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.849724 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.849743 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850042 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850069 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850290 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850402 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850759 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.850782 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.851080 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.851113 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.851272 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.851411 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.853304 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.853817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.854097 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.854186 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.854470 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.854541 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.858883 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.859634 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.861322 4967 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.861745 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.861877 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.861928 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.862113 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.862218 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.862296 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.862319 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.862896 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863184 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863221 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863584 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863696 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863744 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.863804 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864028 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864107 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864493 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864699 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864758 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864594 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864837 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864888 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.864961 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.865073 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.865407 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.865573 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.865586 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.865766 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:44.365733647 +0000 UTC m=+18.951033854 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.870345 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.870370 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.870436 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.870821 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.870862 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.870901 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.871026 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:44.371002631 +0000 UTC m=+18.956302938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871226 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871395 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871513 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871513 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871656 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871744 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871809 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.871843 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872051 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872066 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872412 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872592 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872447 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872748 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.872815 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873002 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873119 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873258 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873268 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873311 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873650 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.873817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.874403 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.875216 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.876564 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.877077 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.877187 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.878017 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.879243 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.879584 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.880485 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.880961 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.881370 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.881529 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.882021 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.882888 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.883905 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.884040 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.884075 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.884397 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.884482 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.885042 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.886201 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.886572 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.888761 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889431 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889457 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889471 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889535 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:44.389515676 +0000 UTC m=+18.974815883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889676 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889692 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889701 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:43 crc kubenswrapper[4967]: E0120 08:49:43.889749 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:44.389739142 +0000 UTC m=+18.975039439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.891469 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.891389 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.891764 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.891884 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.891932 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.892020 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.898694 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.899349 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.899943 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.903573 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.905027 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.908898 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.919373 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.928236 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.937695 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941734 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22fng\" (UniqueName: \"kubernetes.io/projected/73537611-ae3a-4699-a02f-da4f5192d702-kube-api-access-22fng\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941777 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941793 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/73537611-ae3a-4699-a02f-da4f5192d702-hosts-file\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941811 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941856 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941867 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941876 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941886 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941899 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941909 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941918 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941925 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941933 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941941 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941950 4967 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941958 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941967 4967 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.941976 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942001 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942001 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/73537611-ae3a-4699-a02f-da4f5192d702-hosts-file\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942010 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942064 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942067 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942075 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942138 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942153 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942252 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942291 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942307 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942320 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942332 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942344 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942356 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942369 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942382 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942394 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942409 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942421 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942434 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942449 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942462 4967 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942475 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942487 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942501 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942515 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942527 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942539 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942552 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942583 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942596 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942623 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942637 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942650 4967 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942663 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942675 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942689 4967 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942701 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942715 4967 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942728 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942741 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942753 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942766 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942777 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942789 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942801 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942811 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942819 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942830 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942842 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942853 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942866 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942878 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942891 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942902 4967 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942914 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942925 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942937 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942947 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942958 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942969 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942980 4967 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.942991 4967 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943002 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943013 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943024 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943034 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943053 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943065 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943076 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943088 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943101 4967 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943112 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943123 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943565 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943586 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943599 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943668 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943708 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943721 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943735 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943779 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943793 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943806 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943931 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943949 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.943996 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944010 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944022 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944035 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944049 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944061 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944073 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944091 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944103 4967 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944115 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944130 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944142 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944153 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944167 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944179 4967 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944190 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944201 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944213 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944223 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944235 4967 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944246 4967 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944258 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944284 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944322 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944335 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944348 4967 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944360 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944372 4967 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944384 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944395 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944406 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944417 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944428 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944440 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944452 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944464 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944507 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944522 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944535 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944588 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944601 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944932 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944950 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944962 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944977 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.944989 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.946554 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.949287 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.956803 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.957123 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.957825 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22fng\" (UniqueName: \"kubernetes.io/projected/73537611-ae3a-4699-a02f-da4f5192d702-kube-api-access-22fng\") pod \"node-resolver-nxlff\" (UID: \"73537611-ae3a-4699-a02f-da4f5192d702\") " pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.964857 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxlff" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.967673 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.971267 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.975380 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.975632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.975714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.975855 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.975942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:43 crc kubenswrapper[4967]: I0120 08:49:43.976017 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:43Z","lastTransitionTime":"2026-01-20T08:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:43 crc kubenswrapper[4967]: W0120 08:49:43.976164 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73537611_ae3a_4699_a02f_da4f5192d702.slice/crio-4469a4685c73c281286984281482ba4fb56c4ed0f4f6a99b87326141995764ed WatchSource:0}: Error finding container 4469a4685c73c281286984281482ba4fb56c4ed0f4f6a99b87326141995764ed: Status 404 returned error can't find the container with id 4469a4685c73c281286984281482ba4fb56c4ed0f4f6a99b87326141995764ed Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.082173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.082224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.082236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.082357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.082421 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.185033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.185066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.185075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.185088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.185109 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.286874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.286910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.286919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.286937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.286945 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.348719 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:49:45.348682865 +0000 UTC m=+19.933983082 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.348751 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.389078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.389151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.389159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.389173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.389182 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.449862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.449915 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.449942 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.449968 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450096 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450108 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450150 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450165 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450223 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:45.450205025 +0000 UTC m=+20.035505242 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450118 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450285 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450359 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:45.450336479 +0000 UTC m=+20.035636696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450121 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450449 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:45.450429251 +0000 UTC m=+20.035729468 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450122 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: E0120 08:49:44.450503 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:45.450494973 +0000 UTC m=+20.035795190 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.492270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.492300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.492310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.492323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.492332 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.594714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.594756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.594767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.594785 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.594798 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.645383 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 20:06:23.482903561 +0000 UTC Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.697398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.697441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.697453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.697469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.697481 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.794633 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.796456 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.796789 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.797605 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxlff" event={"ID":"73537611-ae3a-4699-a02f-da4f5192d702","Type":"ContainerStarted","Data":"d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.797653 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxlff" event={"ID":"73537611-ae3a-4699-a02f-da4f5192d702","Type":"ContainerStarted","Data":"4469a4685c73c281286984281482ba4fb56c4ed0f4f6a99b87326141995764ed"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.798657 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b2d10e66a2a04091953be78e5455b2705de9d02fc9d053f6d5781437b9bcb60e"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.799200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.799230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.799239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.799252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.799262 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.800637 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.800698 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.800715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"70c0697ae67040a6fb117d5bb03fb3cebb6bcc116aa7a67b98dda19104fd5e97"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.802179 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.802211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4179edf51527ad19c82ba1ff133d95afa28028f9344309749438120a7fdf6244"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.812673 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.824970 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.842703 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.854031 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.872256 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.884308 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.895710 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.902298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.902343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.902355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.902390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.902402 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:44Z","lastTransitionTime":"2026-01-20T08:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.907147 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.920931 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.941437 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.958375 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-bcgrn"] Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.958759 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.960582 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.960766 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.960779 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.961372 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.961374 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6wm4t"] Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.961780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6wm4t" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.961939 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.962221 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.964643 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.964672 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.965808 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.965947 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.967664 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 20 08:49:44 crc kubenswrapper[4967]: I0120 08:49:44.987528 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.000623 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.004402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.004457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.004468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.004485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.004517 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.013176 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.023696 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.038226 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.052179 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054827 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-socket-dir-parent\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054859 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-multus\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054876 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-multus-certs\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054902 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttlgv\" (UniqueName: \"kubernetes.io/projected/5de94293-c4d1-4169-a8df-69843d57278a-kube-api-access-ttlgv\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054918 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-conf-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-os-release\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054949 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-netns\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-etc-kubernetes\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.054996 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-system-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055047 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-bin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055071 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-kubelet\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055093 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgmkx\" (UniqueName: \"kubernetes.io/projected/7ba8630f-92bc-4708-a722-a7e27c747073-kube-api-access-wgmkx\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055157 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5de94293-c4d1-4169-a8df-69843d57278a-rootfs\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055197 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de94293-c4d1-4169-a8df-69843d57278a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055238 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-k8s-cni-cncf-io\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055278 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5de94293-c4d1-4169-a8df-69843d57278a-proxy-tls\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055299 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-cni-binary-copy\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055327 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-cnibin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-hostroot\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.055390 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-multus-daemon-config\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.065225 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.079962 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.106148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.106188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.106201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.106218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.106233 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.120447 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156116 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5de94293-c4d1-4169-a8df-69843d57278a-rootfs\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156182 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de94293-c4d1-4169-a8df-69843d57278a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156203 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156222 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5de94293-c4d1-4169-a8df-69843d57278a-rootfs\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156258 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-k8s-cni-cncf-io\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156297 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-k8s-cni-cncf-io\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156308 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5de94293-c4d1-4169-a8df-69843d57278a-proxy-tls\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156334 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-cni-binary-copy\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156363 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-cnibin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156382 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-hostroot\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156400 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-multus-daemon-config\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156413 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156436 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-socket-dir-parent\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156460 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-multus\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156481 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-multus-certs\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156493 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-hostroot\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttlgv\" (UniqueName: \"kubernetes.io/projected/5de94293-c4d1-4169-a8df-69843d57278a-kube-api-access-ttlgv\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156558 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-conf-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156586 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-os-release\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156607 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-netns\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156602 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-cnibin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156677 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-etc-kubernetes\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156676 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-socket-dir-parent\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156716 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-multus-conf-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156652 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-etc-kubernetes\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156757 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-os-release\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156778 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-multus\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156798 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-system-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156800 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-multus-certs\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156748 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-run-netns\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156838 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-bin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156855 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-system-cni-dir\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156865 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-cni-bin\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156894 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgmkx\" (UniqueName: \"kubernetes.io/projected/7ba8630f-92bc-4708-a722-a7e27c747073-kube-api-access-wgmkx\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156924 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-kubelet\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156961 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-cni-binary-copy\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.156984 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ba8630f-92bc-4708-a722-a7e27c747073-host-var-lib-kubelet\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.157083 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ba8630f-92bc-4708-a722-a7e27c747073-multus-daemon-config\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.157400 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de94293-c4d1-4169-a8df-69843d57278a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.186030 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5de94293-c4d1-4169-a8df-69843d57278a-proxy-tls\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.192263 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttlgv\" (UniqueName: \"kubernetes.io/projected/5de94293-c4d1-4169-a8df-69843d57278a-kube-api-access-ttlgv\") pod \"machine-config-daemon-bcgrn\" (UID: \"5de94293-c4d1-4169-a8df-69843d57278a\") " pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.195787 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgmkx\" (UniqueName: \"kubernetes.io/projected/7ba8630f-92bc-4708-a722-a7e27c747073-kube-api-access-wgmkx\") pod \"multus-6wm4t\" (UID: \"7ba8630f-92bc-4708-a722-a7e27c747073\") " pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.196349 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.208361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.208390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.208398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.208411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.208420 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.212644 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.224465 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.238759 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.249932 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.267874 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.271062 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.277160 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6wm4t" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.299525 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ba8630f_92bc_4708_a722_a7e27c747073.slice/crio-8f511afe689c6139c714c9ddb761ecd3fc98985988c75c8b3d43e5af5150945e WatchSource:0}: Error finding container 8f511afe689c6139c714c9ddb761ecd3fc98985988c75c8b3d43e5af5150945e: Status 404 returned error can't find the container with id 8f511afe689c6139c714c9ddb761ecd3fc98985988c75c8b3d43e5af5150945e Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.317196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.317232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.317243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.317452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.317467 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.354490 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc9bg"] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.355231 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.356649 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-hlcgf"] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.357139 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361037 4967 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361084 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361131 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"env-overrides": failed to list *v1.ConfigMap: configmaps "env-overrides" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361163 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"env-overrides\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"env-overrides\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361136 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361191 4967 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361198 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361211 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361237 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361253 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361252 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361265 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361293 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361302 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-node-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361306 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-config": failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361320 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.361345 4967 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": failed to list *v1.ConfigMap: configmaps "ovnkube-script-lib" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361355 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-script-lib\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.361545 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.361695 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.361681285 +0000 UTC m=+21.946981492 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.378377 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.398465 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420345 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.420340 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.435951 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.450203 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462209 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462252 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-os-release\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462271 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462290 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462308 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-binary-copy\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462323 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462338 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr52w\" (UniqueName: \"kubernetes.io/projected/acb9067c-29d9-4392-8c7e-5198eb5ee93d-kube-api-access-gr52w\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462347 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462358 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462412 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.462393403 +0000 UTC m=+22.047693720 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462419 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462481 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462476 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462526 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462493 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462551 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.462646 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.462625809 +0000 UTC m=+22.047926016 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462696 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462794 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462837 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462865 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462889 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462911 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462964 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.462999 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463036 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cnibin\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463045 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463070 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463078 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463116 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463140 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463197 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463223 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.463174935 +0000 UTC m=+22.048475142 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463259 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49rck\" (UniqueName: \"kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463288 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463317 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463346 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-system-cni-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463388 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.463407 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463605 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.463671 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.463661068 +0000 UTC m=+22.048961515 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.465323 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.479029 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.492305 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.510283 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.522760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.522803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.522813 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.522833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.522846 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.523483 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.541860 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.547886 4967 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548138 4967 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548234 4967 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548311 4967 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548313 4967 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548356 4967 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548377 4967 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548396 4967 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548179 4967 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.548587 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-operator/pods/iptables-alerter-4ln5h/status\": read tcp 38.102.83.136:54918->38.102.83.136:6443: use of closed network connection" Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548734 4967 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: W0120 08:49:45.548776 4967 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564045 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-system-cni-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564099 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564117 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564132 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564171 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-os-release\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564176 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-system-cni-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564215 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564242 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564185 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564264 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564305 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-binary-copy\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564325 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-os-release\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564329 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564351 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564355 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr52w\" (UniqueName: \"kubernetes.io/projected/acb9067c-29d9-4392-8c7e-5198eb5ee93d-kube-api-access-gr52w\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564421 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564449 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564464 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564481 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564503 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564527 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564542 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564629 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564635 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564660 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564664 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564692 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564709 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cnibin\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564718 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564726 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564737 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564765 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564749 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564749 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564789 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564799 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49rck\" (UniqueName: \"kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564819 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564920 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cnibin\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564940 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.564974 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.565064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-binary-copy\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.575965 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.585747 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr52w\" (UniqueName: \"kubernetes.io/projected/acb9067c-29d9-4392-8c7e-5198eb5ee93d-kube-api-access-gr52w\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.590054 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.602342 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.617079 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.625846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.625888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.625899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.625914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.625926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.628961 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.640636 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.646282 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 16:59:37.665165624 +0000 UTC Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.660888 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.675785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.691862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.695925 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.696075 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.696117 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.696183 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.696210 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:45 crc kubenswrapper[4967]: E0120 08:49:45.696330 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.699300 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.700350 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.701769 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.702511 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.703704 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.704315 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.704993 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.706116 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.706940 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.710204 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.710981 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.712486 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.713100 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.713733 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.714797 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.715390 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.718224 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.718815 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.719563 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.721097 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.721578 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.722140 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.722543 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.723189 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.723634 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.724200 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.724827 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.725277 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.725863 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.726392 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.727934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.727958 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.727966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.727979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.727987 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.728744 4967 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.728838 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.729650 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.730365 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.731319 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.732001 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.733452 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.734099 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.737135 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.737804 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.740055 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.740494 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.741505 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.742523 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.743115 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.743628 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.744497 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.745521 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.746230 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.746733 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.747652 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.748142 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.749165 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.749732 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.750188 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.766717 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.779739 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.805674 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerStarted","Data":"650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.806132 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerStarted","Data":"8f511afe689c6139c714c9ddb761ecd3fc98985988c75c8b3d43e5af5150945e"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.807066 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.807090 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.807100 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"f04f0db038122828fa50d03e69bb64767b97f7fc34b88809f3e7368751f72978"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.830175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.830202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.830210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.830224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.830232 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.839156 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.858044 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.869898 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.887417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.900311 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.913496 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.929043 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.932677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.932705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.932716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.932732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.932742 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:45Z","lastTransitionTime":"2026-01-20T08:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.939523 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.953273 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.969970 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:45 crc kubenswrapper[4967]: I0120 08:49:45.987375 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.023743 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.035249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.035289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.035300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.035312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.035321 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.060771 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.106809 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.137761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.138057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.138132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.138209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.138308 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.145655 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.184564 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.221840 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.240894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.240948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.240958 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.240972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.240982 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.264715 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.301302 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.342902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.342939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.342950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.342968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.342980 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.345075 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.385486 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.393903 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.414042 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.438885 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.442490 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.445193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.445252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.445265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.445284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.445297 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.453916 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.483274 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.499207 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.513705 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.547303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.547347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.547361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.547382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.547397 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.553711 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.556050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565324 4967 configmap.go:193] Couldn't get configMap openshift-ovn-kubernetes/ovnkube-script-lib: failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565343 4967 configmap.go:193] Couldn't get configMap openshift-multus/default-cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565357 4967 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-node-metrics-cert: failed to sync secret cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565428 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib podName:9e9f22bb-157c-4556-9066-cd3cdacdf3af nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.06540104 +0000 UTC m=+21.650701287 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovnkube-script-lib" (UniqueName: "kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib") pod "ovnkube-node-fc9bg" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af") : failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565454 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert podName:9e9f22bb-157c-4556-9066-cd3cdacdf3af nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.065441161 +0000 UTC m=+21.650741398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-node-metrics-cert" (UniqueName: "kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert") pod "ovnkube-node-fc9bg" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af") : failed to sync secret cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565481 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist podName:acb9067c-29d9-4392-8c7e-5198eb5ee93d nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.065467952 +0000 UTC m=+21.650768199 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist") pod "multus-additional-cni-plugins-hlcgf" (UID: "acb9067c-29d9-4392-8c7e-5198eb5ee93d") : failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565659 4967 configmap.go:193] Couldn't get configMap openshift-ovn-kubernetes/ovnkube-config: failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.565765 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config podName:9e9f22bb-157c-4556-9066-cd3cdacdf3af nodeName:}" failed. No retries permitted until 2026-01-20 08:49:47.065743919 +0000 UTC m=+21.651044126 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovnkube-config" (UniqueName: "kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config") pod "ovnkube-node-fc9bg" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af") : failed to sync configmap cache: timed out waiting for the condition Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.573238 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.602949 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.614025 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.633772 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.647151 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 14:36:37.77108271 +0000 UTC Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.649432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.649522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.649579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.649696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.649762 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.653755 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.665972 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49rck\" (UniqueName: \"kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.693599 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.714240 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.742807 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.753251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.753434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.753529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.753623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.753707 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.754720 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.774530 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.794541 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.811394 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd"} Jan 20 08:49:46 crc kubenswrapper[4967]: E0120 08:49:46.836956 4967 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.856121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.856168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.856179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.856196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.856214 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.865048 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.873861 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.895116 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.934180 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.953988 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.959753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.959989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.960081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.960166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.960263 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:46Z","lastTransitionTime":"2026-01-20T08:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:46 crc kubenswrapper[4967]: I0120 08:49:46.981584 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:46Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.022345 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.033450 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.062807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.062846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.062871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.062891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.062902 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.079318 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.080038 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.080112 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.080218 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.080265 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.080967 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.081344 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/acb9067c-29d9-4392-8c7e-5198eb5ee93d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hlcgf\" (UID: \"acb9067c-29d9-4392-8c7e-5198eb5ee93d\") " pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.083655 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") pod \"ovnkube-node-fc9bg\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.085735 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.124604 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.163845 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.165809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.165866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.165882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.165905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.165920 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.171951 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.180964 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" Jan 20 08:49:47 crc kubenswrapper[4967]: W0120 08:49:47.188178 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9f22bb_157c_4556_9066_cd3cdacdf3af.slice/crio-b32022fc51bdc7093b03b3236f1bb22dea985de51fe17320b6663cdcd3bf6564 WatchSource:0}: Error finding container b32022fc51bdc7093b03b3236f1bb22dea985de51fe17320b6663cdcd3bf6564: Status 404 returned error can't find the container with id b32022fc51bdc7093b03b3236f1bb22dea985de51fe17320b6663cdcd3bf6564 Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.207820 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.245302 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.268818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.268857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.268867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.268884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.268895 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.285129 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.322977 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.363717 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.371573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.371623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.371632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.371652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.371662 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.383683 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.383910 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:49:51.383892093 +0000 UTC m=+25.969192310 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.403851 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.441750 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.474016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.474052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.474061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.474081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.474106 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.481814 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.485463 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.485532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.485582 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485633 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.485651 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485704 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:51.48568538 +0000 UTC m=+26.070985587 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485766 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485816 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485855 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485877 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485825 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:51.485808304 +0000 UTC m=+26.071108521 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.485817 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.486004 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:51.485982099 +0000 UTC m=+26.071282356 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.486021 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.486032 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.486057 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:51.486049821 +0000 UTC m=+26.071350028 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.527219 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.565283 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.576933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.577336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.577354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.577378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.577395 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.604448 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.647781 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 19:01:49.789838979 +0000 UTC Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.665159 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.679987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.680024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.680037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.680052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.680063 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.687532 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.694077 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.694112 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.694077 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.694205 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.694390 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:47 crc kubenswrapper[4967]: E0120 08:49:47.694476 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.721704 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.762011 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.782658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.782705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.782718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.782732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.782741 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.802858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.815145 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb" exitCode=0 Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.815228 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.815259 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"b32022fc51bdc7093b03b3236f1bb22dea985de51fe17320b6663cdcd3bf6564"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.816864 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea" exitCode=0 Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.816946 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.817003 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerStarted","Data":"c086c8b2a26094e47ea7272575e25a0d84ed6edca44cd21ac5da0bf7ee1ae2bf"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.842736 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.886248 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.887971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.888002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.888012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.888028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.888040 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.927059 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.960867 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:47Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.996218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.996247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.996258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.996276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:47 crc kubenswrapper[4967]: I0120 08:49:47.996287 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:47Z","lastTransitionTime":"2026-01-20T08:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.006301 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.040264 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.083208 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.098204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.098251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.098265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.098282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.098293 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.126403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.167772 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.200241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.200282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.200293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.200309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.200320 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.203203 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.243808 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.286000 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.302437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.302468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.302477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.302490 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.302503 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.321473 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.361812 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.402391 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.404783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.404820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.404830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.404847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.404858 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.441282 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.507870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.507906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.507917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.507933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.507942 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.609842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.609897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.609914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.609940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.609957 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.648512 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 08:56:08.646386782 +0000 UTC Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.711832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.711858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.711866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.711878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.711886 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.744552 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-h8vz7"] Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.745209 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.748045 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.748878 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.749771 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.749795 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.757593 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.770038 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.794142 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.805056 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-host\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.805138 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-serviceca\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.805254 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlch\" (UniqueName: \"kubernetes.io/projected/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-kube-api-access-ltlch\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.811926 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.814158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.814184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.814196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.814212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.814225 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824636 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824685 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824701 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824713 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824725 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.824738 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.826250 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8" exitCode=0 Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.826290 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.830325 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.846548 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.859350 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.875148 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.888001 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.906273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-host\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.906354 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-serviceca\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.906393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlch\" (UniqueName: \"kubernetes.io/projected/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-kube-api-access-ltlch\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.906435 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-host\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.907690 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-serviceca\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.916876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.916928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.916942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.916962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.916973 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:48Z","lastTransitionTime":"2026-01-20T08:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.924923 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.960207 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlch\" (UniqueName: \"kubernetes.io/projected/4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2-kube-api-access-ltlch\") pod \"node-ca-h8vz7\" (UID: \"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\") " pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:48 crc kubenswrapper[4967]: I0120 08:49:48.985734 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:48Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.019318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.019375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.019392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.019417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.019435 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.022583 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.060813 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8vz7" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.060882 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.103455 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.122828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.122894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.122915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.122938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.122955 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.141146 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: W0120 08:49:49.143511 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c2fbe33_7bc2_451f_9c86_bbf2bb8f37b2.slice/crio-dd026dc26e14052e47611467cb75ccb79ce03bd6e92aa4bac83eaab38a665a61 WatchSource:0}: Error finding container dd026dc26e14052e47611467cb75ccb79ce03bd6e92aa4bac83eaab38a665a61: Status 404 returned error can't find the container with id dd026dc26e14052e47611467cb75ccb79ce03bd6e92aa4bac83eaab38a665a61 Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.182740 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.223283 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.225384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.225414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.225425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.225441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.225452 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.262137 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.302020 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.327996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.328039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.328055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.328077 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.328094 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.343456 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.383253 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.430573 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.431336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.431379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.431394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.431416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.431432 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.467421 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.504524 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.534296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.534366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.534390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.534418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.534441 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.544983 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.588209 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.624909 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.638573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.638668 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.638688 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.638714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.638734 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.649298 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 15:02:38.337680504 +0000 UTC Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.668854 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.694088 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.694198 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:49 crc kubenswrapper[4967]: E0120 08:49:49.694252 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:49 crc kubenswrapper[4967]: E0120 08:49:49.694381 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.694452 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:49 crc kubenswrapper[4967]: E0120 08:49:49.694572 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.741774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.741840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.741861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.741888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.741909 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.831607 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8vz7" event={"ID":"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2","Type":"ContainerStarted","Data":"e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.831728 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8vz7" event={"ID":"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2","Type":"ContainerStarted","Data":"dd026dc26e14052e47611467cb75ccb79ce03bd6e92aa4bac83eaab38a665a61"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.834579 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73" exitCode=0 Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.834653 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.845028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.845100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.845126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.845156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.845178 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.855036 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.872593 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.897192 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.917509 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.932404 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.944571 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.948870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.948898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.948906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.948920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.948946 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:49Z","lastTransitionTime":"2026-01-20T08:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.956401 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:49 crc kubenswrapper[4967]: I0120 08:49:49.980214 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:49Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.020719 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.050985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.051010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.051018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.051032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.051039 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.067235 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.104480 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.142020 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.153601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.153637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.153647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.153660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.153670 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.183447 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.221852 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.255991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.256028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.256041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.256056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.256067 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.263735 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.303311 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.344183 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.359851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.359893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.359919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.359940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.359957 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.385764 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.423321 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.460239 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.462061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.462093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.462106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.462122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.462134 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.506300 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.543342 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.564282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.564324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.564335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.564354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.564365 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.581951 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.622878 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.649578 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 06:39:21.379200545 +0000 UTC Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.660733 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.666753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.666796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.666810 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.666829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.666842 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.704550 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.769856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.769933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.769945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.769962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.769974 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.783943 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.807857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.840020 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e" exitCode=0 Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.840088 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.844553 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.856716 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.869334 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.871786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.871833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.871846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.871863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.871876 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.908544 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.942885 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.976179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.976213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.976228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.976246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.976259 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:50Z","lastTransitionTime":"2026-01-20T08:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:50 crc kubenswrapper[4967]: I0120 08:49:50.981661 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:50Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.023232 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.065800 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.078085 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.078110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.078117 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.078130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.078145 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.100220 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.142173 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.181710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.181748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.181759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.181776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.181789 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.190025 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.223492 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.265216 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.283807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.283843 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.286855 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.286927 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.286940 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.303889 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.344432 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.389441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.389475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.389484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.389497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.389506 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.429237 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.429453 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.429419468 +0000 UTC m=+34.014719705 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.491846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.491879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.491888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.491902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.491911 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.530622 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.530687 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.530714 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.530745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.530875 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.530896 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.530909 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.530959 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.530943389 +0000 UTC m=+34.116243606 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531387 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531429 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.531418402 +0000 UTC m=+34.116718619 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531465 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531489 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.531481404 +0000 UTC m=+34.116781631 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531594 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531909 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531940 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.531979 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.531966667 +0000 UTC m=+34.117266884 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.594277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.594333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.594348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.594366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.594378 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.650525 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 19:47:14.570883442 +0000 UTC Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.694314 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.694346 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.694396 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.694520 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.694588 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:51 crc kubenswrapper[4967]: E0120 08:49:51.694679 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.696366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.696562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.696584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.696635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.696648 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.799648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.799698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.799709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.799729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.799740 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.850318 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca" exitCode=0 Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.850362 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.869545 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.879647 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.890943 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.902343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.902376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.902387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.902413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.902425 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:51Z","lastTransitionTime":"2026-01-20T08:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.912628 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.928364 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.943049 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.955772 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.968342 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.982040 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:51 crc kubenswrapper[4967]: I0120 08:49:51.994295 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:51Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.004718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.004759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.004771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.004786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.004796 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.009192 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.022943 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.034049 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.049176 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.107959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.108020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.108038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.108062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.108081 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.210217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.210256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.210269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.210285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.210296 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.312586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.312645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.312656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.312671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.312680 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.416204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.416268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.416288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.416310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.416325 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.519379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.519422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.519445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.519464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.519474 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.621574 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.621627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.621638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.621653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.621664 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.651028 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 08:59:16.301436778 +0000 UTC Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.724933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.725048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.725084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.725111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.725132 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.828370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.828422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.828431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.828450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.828464 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.859262 4967 generic.go:334] "Generic (PLEG): container finished" podID="acb9067c-29d9-4392-8c7e-5198eb5ee93d" containerID="670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12" exitCode=0 Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.859343 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerDied","Data":"670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.885203 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.908017 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.919364 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.931309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.931343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.931357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.931376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.931390 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:52Z","lastTransitionTime":"2026-01-20T08:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.936226 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.946361 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.955998 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.977107 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:52 crc kubenswrapper[4967]: I0120 08:49:52.989248 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:52Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.002483 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.017166 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034208 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.034998 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.048535 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.058904 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.072384 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.135657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.135682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.135692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.135709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.135721 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.237695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.237731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.237742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.237758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.237768 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.340012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.340074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.340098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.340120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.340134 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.442382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.442442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.442453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.442472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.442483 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.544759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.544807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.544815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.544830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.544839 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.647800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.647851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.647862 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.647882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.647897 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.652052 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 21:45:58.399234264 +0000 UTC Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.694024 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.694065 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.694387 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.694500 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.694065 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.694757 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.749962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.750005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.750019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.750036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.750050 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.853027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.853110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.853135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.853164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.853187 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.869024 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" event={"ID":"acb9067c-29d9-4392-8c7e-5198eb5ee93d","Type":"ContainerStarted","Data":"dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.876865 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.877899 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.877958 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.886066 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.904117 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.912116 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.917764 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.926188 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.940411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.940463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.940479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.940505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.940522 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.945964 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.955429 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.959977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.960060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.960070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.960084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.960096 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.960921 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.970872 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.974888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.974927 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.974943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.974957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.974967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.975752 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: E0120 08:49:53.986857 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.991285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.991332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.991350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.991372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.991389 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:53Z","lastTransitionTime":"2026-01-20T08:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:53 crc kubenswrapper[4967]: I0120 08:49:53.996441 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:53Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: E0120 08:49:54.005185 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.009389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.009644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.009743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.009853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.009949 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.012932 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: E0120 08:49:54.022163 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: E0120 08:49:54.022518 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.024324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.024415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.024439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.024464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.024481 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.027650 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.039890 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.052858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.066326 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.080424 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.094869 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.106508 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.117758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.127143 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.127171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.127181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.127198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.127209 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.134163 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.147155 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.158725 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.171433 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.186328 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.200858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.230422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.230482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.230502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.230531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.230548 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.241388 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.262299 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.276761 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.290415 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.311786 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.326136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:54Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.333914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.334025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.334050 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.334081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.334104 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.436647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.436691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.436702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.436721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.436733 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.538972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.539020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.539033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.539053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.539070 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.641684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.641903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.641978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.642052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.642118 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.652978 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 23:12:01.941264557 +0000 UTC Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.744367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.744409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.744420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.744435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.744447 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.846527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.846567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.846577 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.846590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.846598 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.880323 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.948390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.948682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.948692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.948706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:54 crc kubenswrapper[4967]: I0120 08:49:54.948714 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:54Z","lastTransitionTime":"2026-01-20T08:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.051325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.051347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.051355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.051367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.051376 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.153851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.153921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.153932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.153950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.153961 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.256838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.256878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.256895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.256919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.256935 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.359331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.359401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.359424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.359488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.359512 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.461948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.461985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.461995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.462012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.462023 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.564950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.564989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.564998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.565012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.565022 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.653585 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 19:44:48.244343106 +0000 UTC Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.667257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.667288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.667298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.667312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.667321 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.694107 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.694163 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.694274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:55 crc kubenswrapper[4967]: E0120 08:49:55.694390 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:55 crc kubenswrapper[4967]: E0120 08:49:55.695520 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:55 crc kubenswrapper[4967]: E0120 08:49:55.695654 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.714872 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.729724 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.742444 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.762727 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.771898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.771943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.771955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.771972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.771983 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.782049 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.795987 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.810934 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.826026 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.837670 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.851070 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.861692 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.871076 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.873820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.873845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.873856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.873873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.873889 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.885902 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/0.log" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.886563 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.889348 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a" exitCode=1 Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.889393 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.890428 4967 scope.go:117] "RemoveContainer" containerID="48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.898600 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.909284 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.919048 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.932026 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.948735 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.962143 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975186 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.975644 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:55Z","lastTransitionTime":"2026-01-20T08:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.987121 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:55 crc kubenswrapper[4967]: I0120 08:49:55.998440 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.011091 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.025999 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.036703 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.048739 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.060562 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.073323 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.079343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.079382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.079396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.079414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.079427 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.181594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.181663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.181678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.181695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.181707 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.284956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.285007 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.285018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.285036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.285051 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.387761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.387805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.387817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.387835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.387849 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.491025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.491086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.491099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.491119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.491138 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.593653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.593719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.593729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.593745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.593754 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.654658 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 05:25:52.625910363 +0000 UTC Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.695487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.695529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.695541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.695557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.695568 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.798739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.798789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.798802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.798822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.798835 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.894216 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/0.log" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.897156 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.897295 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.901491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.901521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.901531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.901546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.901565 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:56Z","lastTransitionTime":"2026-01-20T08:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.913605 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.928099 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.942787 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.955144 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.967429 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.983378 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:56 crc kubenswrapper[4967]: I0120 08:49:56.997358 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:56Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.003264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.003307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.003318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.003336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.003348 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.024064 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.048594 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.071790 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.083203 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.094573 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.105547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.105590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.105605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.105645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.105660 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.106199 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.117256 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.207298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.207336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.207347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.207361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.207371 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.309572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.309642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.309654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.309670 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.309682 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.411999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.412082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.412106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.412136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.412159 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.514776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.514845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.514870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.514900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.514921 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.555062 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x"] Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.555985 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.558524 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.560060 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.579639 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.600200 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.617090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.617129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.617138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.617152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.617162 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.619387 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.639939 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.657313 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 21:06:56.151596861 +0000 UTC Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.664481 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.675480 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.693313 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.693343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.693385 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:57 crc kubenswrapper[4967]: E0120 08:49:57.693456 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:57 crc kubenswrapper[4967]: E0120 08:49:57.693569 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:57 crc kubenswrapper[4967]: E0120 08:49:57.693666 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.698447 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.698586 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhd48\" (UniqueName: \"kubernetes.io/projected/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-kube-api-access-xhd48\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.698727 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.698822 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.699378 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.717271 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.719407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.719463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.719481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.719507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.719527 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.730806 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.746582 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.770177 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.796007 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799210 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799257 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799280 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhd48\" (UniqueName: \"kubernetes.io/projected/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-kube-api-access-xhd48\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799299 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.799943 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.807685 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.809801 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.822936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.823011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.823035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.823067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.823091 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.823177 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhd48\" (UniqueName: \"kubernetes.io/projected/e5c4e15f-99b5-46ed-93bf-15b3413e6eab-kube-api-access-xhd48\") pod \"ovnkube-control-plane-749d76644c-mfb5x\" (UID: \"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.825918 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.840334 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.880009 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" Jan 20 08:49:57 crc kubenswrapper[4967]: W0120 08:49:57.895293 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5c4e15f_99b5_46ed_93bf_15b3413e6eab.slice/crio-0cf667f9ba67f642858c3c94276eba2e242358244dfb597429b9b1ebb5de7e79 WatchSource:0}: Error finding container 0cf667f9ba67f642858c3c94276eba2e242358244dfb597429b9b1ebb5de7e79: Status 404 returned error can't find the container with id 0cf667f9ba67f642858c3c94276eba2e242358244dfb597429b9b1ebb5de7e79 Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.902632 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/1.log" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.903219 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/0.log" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.906280 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" exitCode=1 Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.906368 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.906454 4967 scope.go:117] "RemoveContainer" containerID="48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.906894 4967 scope.go:117] "RemoveContainer" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" Jan 20 08:49:57 crc kubenswrapper[4967]: E0120 08:49:57.907065 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.908014 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" event={"ID":"e5c4e15f-99b5-46ed-93bf-15b3413e6eab","Type":"ContainerStarted","Data":"0cf667f9ba67f642858c3c94276eba2e242358244dfb597429b9b1ebb5de7e79"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.925968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.926000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.926010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.926026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.926036 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:57Z","lastTransitionTime":"2026-01-20T08:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.926692 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.941368 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.952068 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.962229 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.973133 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.986410 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:49:57 crc kubenswrapper[4967]: I0120 08:49:57.990841 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:57Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.003461 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.016344 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.027005 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.028378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.028416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.028427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.028442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.028454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.037728 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.049682 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.060905 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.071502 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.080210 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.093078 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.105158 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.120033 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.130976 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.131917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.131979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.131996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.132022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.132034 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.142254 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.152700 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.160883 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.173039 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.188140 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.201999 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.214505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.228095 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.234269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.234304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.234317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.234335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.234349 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.243160 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.255785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.272412 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.285303 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lxp6p"] Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.285694 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: E0120 08:49:58.285751 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.289596 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.301786 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.314008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.326166 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.336796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.336826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.336835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.336848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.336856 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.339428 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.350150 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.360839 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.385941 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.396505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.405232 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.405323 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq2xq\" (UniqueName: \"kubernetes.io/projected/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-kube-api-access-gq2xq\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.408512 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.421114 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.433866 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.438887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.438922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.438931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.438946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.438960 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.447060 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.456484 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.465708 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.477902 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.491201 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.506897 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq2xq\" (UniqueName: \"kubernetes.io/projected/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-kube-api-access-gq2xq\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.507016 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: E0120 08:49:58.507185 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:49:58 crc kubenswrapper[4967]: E0120 08:49:58.507268 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:49:59.007245103 +0000 UTC m=+33.592545340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.523588 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq2xq\" (UniqueName: \"kubernetes.io/projected/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-kube-api-access-gq2xq\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.541527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.541584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.541600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.541646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.541664 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.645333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.645430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.645452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.645480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.645500 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.657865 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 20:11:32.425349627 +0000 UTC Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.748150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.748193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.748211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.748230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.748243 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.850853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.850925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.850941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.850967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.850985 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.914176 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" event={"ID":"e5c4e15f-99b5-46ed-93bf-15b3413e6eab","Type":"ContainerStarted","Data":"b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.914490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" event={"ID":"e5c4e15f-99b5-46ed-93bf-15b3413e6eab","Type":"ContainerStarted","Data":"8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.917155 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/1.log" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.927337 4967 scope.go:117] "RemoveContainer" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" Jan 20 08:49:58 crc kubenswrapper[4967]: E0120 08:49:58.927527 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.940496 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48eae1d3ba965b72b25a45bc0bc4ca0f63ff92a0038bad08c9b731dcfed88a6a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:55Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.058705 6235 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0120 08:49:55.058827 6235 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:49:55.059081 6235 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0120 08:49:55.059135 6235 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059200 6235 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:49:55.059149 6235 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0120 08:49:55.059230 6235 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:49:55.059260 6235 factory.go:656] Stopping watch factory\\\\nI0120 08:49:55.059275 6235 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0120 08:49:55.059297 6235 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:49:55.059334 6235 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.953520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.953574 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.953589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.953640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.953658 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:58Z","lastTransitionTime":"2026-01-20T08:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.956363 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.969143 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.982016 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:58 crc kubenswrapper[4967]: I0120 08:49:58.995875 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:58Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.006525 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.013056 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.013297 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.013450 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:00.013423795 +0000 UTC m=+34.598724032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.016418 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.030239 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.043661 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.055961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.056107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.056193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.056350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.056444 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.059045 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.072051 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.088399 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.104569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.115181 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.132929 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.147421 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.158924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.158972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.158988 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.159010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.159027 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.162596 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.178653 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.191126 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.203464 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.214909 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.224907 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.239403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.249882 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.261640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.261683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.261694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.261711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.261723 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.267750 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.295421 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.316909 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.332521 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.345000 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.363231 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.365451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.365482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.365491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.365507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.365518 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.373991 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.383729 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:49:59Z is after 2025-08-24T17:21:41Z" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.467936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.468295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.468312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.468336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.468355 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.517973 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.518213 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:50:15.518194678 +0000 UTC m=+50.103494885 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.570822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.570890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.570913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.570944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.570968 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.618577 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.618673 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.618705 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.618753 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618764 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618864 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:15.618844174 +0000 UTC m=+50.204144381 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618866 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618887 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618877 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618961 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.619008 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.619025 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:15.618986538 +0000 UTC m=+50.204286805 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.618900 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.619032 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.619157 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:15.619123032 +0000 UTC m=+50.204423279 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.619191 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:15.619176373 +0000 UTC m=+50.204476580 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.658391 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 13:19:33.637884946 +0000 UTC Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.673808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.673886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.673912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.673943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.673967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.693503 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.693598 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.693691 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.693812 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.693831 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.693988 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.694344 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:49:59 crc kubenswrapper[4967]: E0120 08:49:59.694405 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.777411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.777488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.777506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.777528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.777545 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.880197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.880240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.880252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.880269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.880281 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.982582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.982687 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.982710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.982740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:49:59 crc kubenswrapper[4967]: I0120 08:49:59.982765 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:49:59Z","lastTransitionTime":"2026-01-20T08:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.023812 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:00 crc kubenswrapper[4967]: E0120 08:50:00.024098 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:00 crc kubenswrapper[4967]: E0120 08:50:00.024330 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:02.024232205 +0000 UTC m=+36.609532582 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.085837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.085892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.085908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.085928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.085942 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.188932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.188972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.188981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.188997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.189006 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.291506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.291538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.291546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.291559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.291570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.393969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.394010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.394020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.394057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.394075 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.495911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.495943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.495953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.495966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.495976 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.597842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.597882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.597894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.597908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.597921 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.659491 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 10:04:51.952448281 +0000 UTC Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.675991 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.676887 4967 scope.go:117] "RemoveContainer" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" Jan 20 08:50:00 crc kubenswrapper[4967]: E0120 08:50:00.677067 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.700340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.700380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.700393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.700410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.700423 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.802118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.802192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.802217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.802247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.802267 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.904763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.904793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.904800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.904812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:00 crc kubenswrapper[4967]: I0120 08:50:00.904821 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:00Z","lastTransitionTime":"2026-01-20T08:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.007132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.007220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.007236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.007257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.007272 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.110720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.110781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.110793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.110815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.110829 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.214302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.214684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.214827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.214961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.215087 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.318393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.318465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.318488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.318520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.318544 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.422472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.422524 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.422540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.422562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.422580 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.526232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.526289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.526306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.526334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.526350 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.629070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.629319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.629328 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.629340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.629352 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.660138 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 08:01:57.26680985 +0000 UTC Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.694177 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.694267 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.694187 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:01 crc kubenswrapper[4967]: E0120 08:50:01.694345 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:01 crc kubenswrapper[4967]: E0120 08:50:01.694440 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.694281 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:01 crc kubenswrapper[4967]: E0120 08:50:01.694629 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:01 crc kubenswrapper[4967]: E0120 08:50:01.694801 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.732304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.732348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.732361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.732377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.732389 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.835266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.835336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.835347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.835360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.835371 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.937918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.937965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.937978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.937997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:01 crc kubenswrapper[4967]: I0120 08:50:01.938008 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:01Z","lastTransitionTime":"2026-01-20T08:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.040957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.041048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.041070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.041101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.041118 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.048687 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:02 crc kubenswrapper[4967]: E0120 08:50:02.048870 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:02 crc kubenswrapper[4967]: E0120 08:50:02.048960 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:06.04893481 +0000 UTC m=+40.634235057 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.144542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.144592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.144645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.144673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.144710 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.247915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.247965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.247979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.247998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.248017 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.350323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.350361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.350391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.350404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.350413 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.453086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.453151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.453174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.453199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.453217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.556326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.556392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.556415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.556444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.556471 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.659324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.659377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.659394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.659416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.659432 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.660642 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 19:02:50.84432589 +0000 UTC Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.762047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.762100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.762118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.762142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.762159 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.864990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.865331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.865361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.865402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.865419 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.967101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.967149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.967161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.967180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:02 crc kubenswrapper[4967]: I0120 08:50:02.967193 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:02Z","lastTransitionTime":"2026-01-20T08:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.069735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.070053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.070163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.070253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.070332 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.173160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.173460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.173530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.173602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.173702 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.275673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.275715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.275723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.275736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.275744 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.378737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.378812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.378837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.378866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.378888 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.533151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.533435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.533499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.533558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.533655 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.636072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.636287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.636367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.636450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.636528 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.661135 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 09:18:27.425611131 +0000 UTC Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.693699 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.693779 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:03 crc kubenswrapper[4967]: E0120 08:50:03.693926 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.693982 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.694015 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:03 crc kubenswrapper[4967]: E0120 08:50:03.694091 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:03 crc kubenswrapper[4967]: E0120 08:50:03.694175 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:03 crc kubenswrapper[4967]: E0120 08:50:03.694377 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.739168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.739229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.739252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.739282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.739305 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.841986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.842020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.842033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.842050 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.842063 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.944923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.945298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.945486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.945723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:03 crc kubenswrapper[4967]: I0120 08:50:03.945926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:03Z","lastTransitionTime":"2026-01-20T08:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.048188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.048223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.048231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.048245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.048255 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.150856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.150905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.150918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.150933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.150943 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.253678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.253713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.253724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.253741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.253753 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.357110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.357170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.357192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.357223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.357246 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.361817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.361865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.361886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.361909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.361931 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.381377 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:04Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.386566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.386618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.386628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.386644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.386654 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.400487 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:04Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.405018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.405069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.405080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.405098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.405110 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.418398 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:04Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.423184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.423243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.423306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.423330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.423347 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.435159 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:04Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.438080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.438126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.438154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.438168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.438179 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.452114 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:04Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:04 crc kubenswrapper[4967]: E0120 08:50:04.452226 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.459542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.459579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.459588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.459602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.459666 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.562428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.562473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.562482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.562495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.562505 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.662088 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 08:50:47.552551694 +0000 UTC Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.665565 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.665666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.665685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.665711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.665728 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.768983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.769046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.769070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.769094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.769110 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.871694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.871952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.872030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.872109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.872198 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.974858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.974927 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.974961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.974985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:04 crc kubenswrapper[4967]: I0120 08:50:04.975001 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:04Z","lastTransitionTime":"2026-01-20T08:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.077374 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.077414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.077424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.077437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.077446 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.180383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.180417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.180426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.180440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.180451 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.283371 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.283431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.283449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.283473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.283490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.386803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.386970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.386990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.387006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.387016 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.489959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.490012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.490029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.490088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.490106 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.592406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.592469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.592487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.592511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.592528 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.663354 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 18:43:42.976249805 +0000 UTC Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.693961 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:05 crc kubenswrapper[4967]: E0120 08:50:05.694091 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.694126 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.694245 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:05 crc kubenswrapper[4967]: E0120 08:50:05.694353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.693986 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:05 crc kubenswrapper[4967]: E0120 08:50:05.694474 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:05 crc kubenswrapper[4967]: E0120 08:50:05.694653 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.695404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.695427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.695435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.695448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.695457 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.712023 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.722317 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.732640 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.772621 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.797738 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.798033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.798053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.798060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.798072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.798082 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.809862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.821108 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.831721 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.841317 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.854239 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.866694 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.877726 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.888574 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.897809 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.900269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.900301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.900327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.900340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.900349 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:05Z","lastTransitionTime":"2026-01-20T08:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.911019 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:05 crc kubenswrapper[4967]: I0120 08:50:05.921981 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:05Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.003079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.003684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.003714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.003740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.003751 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.086761 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:06 crc kubenswrapper[4967]: E0120 08:50:06.087041 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:06 crc kubenswrapper[4967]: E0120 08:50:06.087165 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:14.087137006 +0000 UTC m=+48.672437243 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.106317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.106358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.106368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.106381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.106393 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.209716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.209793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.209816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.209845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.209894 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.312054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.312110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.312129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.312154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.312172 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.414578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.414709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.414743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.414794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.414819 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.517730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.517787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.517804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.517827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.517843 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.620814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.621021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.621090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.621210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.621290 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.664427 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 00:21:48.513667137 +0000 UTC Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.724122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.724184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.724198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.724219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.724232 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.826443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.826482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.826491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.826508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.826517 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.928882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.928918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.928926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.928938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:06 crc kubenswrapper[4967]: I0120 08:50:06.928948 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:06Z","lastTransitionTime":"2026-01-20T08:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.031059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.031118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.031134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.031153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.031168 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.134534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.134590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.134606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.134677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.134699 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.237387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.237426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.237440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.237463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.237479 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.340588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.340674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.340687 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.340735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.340748 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.443835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.443884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.443900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.443923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.443942 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.547196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.547260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.547279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.547304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.547324 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.650547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.650652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.650672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.650696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.650713 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.665331 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 06:33:22.534953542 +0000 UTC Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.693795 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.694184 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:07 crc kubenswrapper[4967]: E0120 08:50:07.694478 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.694896 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.694941 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:07 crc kubenswrapper[4967]: E0120 08:50:07.695108 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:07 crc kubenswrapper[4967]: E0120 08:50:07.695370 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:07 crc kubenswrapper[4967]: E0120 08:50:07.695605 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.753494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.753550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.753567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.753593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.753644 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.856243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.856487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.856550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.856666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.856727 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.972440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.972511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.972532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.972559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:07 crc kubenswrapper[4967]: I0120 08:50:07.972579 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:07Z","lastTransitionTime":"2026-01-20T08:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.075487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.075532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.075549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.075572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.075591 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.178753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.179018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.179157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.179288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.179447 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.282946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.283034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.283067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.283098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.283124 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.387199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.387263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.387281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.387306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.387327 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.490321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.490378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.490393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.490416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.490433 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.592717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.592759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.592769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.592793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.592804 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.665683 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 09:05:25.05868003 +0000 UTC Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.695776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.695872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.695891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.695914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.695933 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.799423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.799492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.799514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.799782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.799805 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.901993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.902067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.902090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.902124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:08 crc kubenswrapper[4967]: I0120 08:50:08.902148 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:08Z","lastTransitionTime":"2026-01-20T08:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.005485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.005530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.005541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.005559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.005570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.108923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.109011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.109029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.109055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.109084 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.211966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.212025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.212044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.212068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.212090 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.314716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.314790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.314815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.314846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.314867 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.418108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.418171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.418187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.418210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.418226 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.521359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.521420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.521437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.521460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.521476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.623535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.623582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.623594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.623635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.623648 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.666074 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 07:20:06.207660006 +0000 UTC Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.693841 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.693895 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.693918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:09 crc kubenswrapper[4967]: E0120 08:50:09.694123 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.694168 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:09 crc kubenswrapper[4967]: E0120 08:50:09.694361 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:09 crc kubenswrapper[4967]: E0120 08:50:09.694475 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:09 crc kubenswrapper[4967]: E0120 08:50:09.694662 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.725974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.726029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.726048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.726071 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.726088 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.829245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.829311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.829333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.829363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.829384 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.933077 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.933159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.933183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.933212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:09 crc kubenswrapper[4967]: I0120 08:50:09.933235 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:09Z","lastTransitionTime":"2026-01-20T08:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.035968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.036011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.036027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.036044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.036055 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.138807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.138879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.138896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.138922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.138940 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.242606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.242724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.242750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.242782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.242804 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.346448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.346522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.346539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.346564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.346581 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.449340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.449401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.449414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.449438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.449450 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.551871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.551934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.551951 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.551976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.551993 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.655365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.655414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.655430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.655456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.655474 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.666485 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 06:06:02.929064605 +0000 UTC Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.759523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.759598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.759649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.759681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.759699 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.862350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.862424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.862445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.862475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.862494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.965641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.965697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.965713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.965737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:10 crc kubenswrapper[4967]: I0120 08:50:10.965755 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:10Z","lastTransitionTime":"2026-01-20T08:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.068226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.068280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.068296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.068317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.068332 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.174140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.174208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.174225 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.174250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.174267 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.276514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.276566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.276583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.276643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.276662 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.380394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.380456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.380478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.380505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.380528 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.484196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.484283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.484308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.484341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.484367 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.587552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.587648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.587675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.587707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.587731 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.667581 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 02:32:25.329299771 +0000 UTC Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.691078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.691134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.691152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.691180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.691196 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.693403 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.693478 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:11 crc kubenswrapper[4967]: E0120 08:50:11.693690 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.693746 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.693808 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:11 crc kubenswrapper[4967]: E0120 08:50:11.693973 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:11 crc kubenswrapper[4967]: E0120 08:50:11.694114 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:11 crc kubenswrapper[4967]: E0120 08:50:11.694384 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.795701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.795790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.795816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.795849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.795884 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.899148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.899194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.899204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.899220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:11 crc kubenswrapper[4967]: I0120 08:50:11.899228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:11Z","lastTransitionTime":"2026-01-20T08:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.002974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.003061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.003086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.003118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.003437 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.106704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.106766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.106786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.106811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.106831 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.209470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.209525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.209542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.209563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.209579 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.312576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.312672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.312685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.312703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.312716 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.415283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.415321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.415332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.415347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.415357 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.517549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.517642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.517660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.517686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.517704 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.620795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.620858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.620876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.620900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.620917 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.668671 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 13:21:08.470998463 +0000 UTC Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.722958 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.722985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.722994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.723008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.723016 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.826121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.826207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.826234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.826270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.826294 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.929647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.929690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.929701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.929725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:12 crc kubenswrapper[4967]: I0120 08:50:12.929737 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:12Z","lastTransitionTime":"2026-01-20T08:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.032333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.032380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.032391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.032408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.032419 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.135351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.135390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.135400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.135415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.135424 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.237917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.238010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.238035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.238073 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.238097 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.341064 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.341146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.341165 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.341194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.341213 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.445535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.445600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.445656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.445681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.445698 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.548844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.548915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.548939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.548964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.548983 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.652049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.652123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.652145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.652179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.652200 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.669805 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 12:34:14.538602876 +0000 UTC Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.693700 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.693774 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:13 crc kubenswrapper[4967]: E0120 08:50:13.693844 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.693925 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:13 crc kubenswrapper[4967]: E0120 08:50:13.693939 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.693985 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:13 crc kubenswrapper[4967]: E0120 08:50:13.694094 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:13 crc kubenswrapper[4967]: E0120 08:50:13.694232 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.695751 4967 scope.go:117] "RemoveContainer" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.755832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.756002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.756109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.756220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.756321 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.756080 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.769026 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.776411 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.790797 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.813523 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.828698 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.839842 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.853989 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.861896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.861933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.861943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.861961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.861974 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.878459 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.902430 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.923301 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.935552 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.947576 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.962425 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.964057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.964085 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.964094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.964106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.964117 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:13Z","lastTransitionTime":"2026-01-20T08:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.975647 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.981486 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/1.log" Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.984968 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721"} Jan 20 08:50:13 crc kubenswrapper[4967]: I0120 08:50:13.993205 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:13Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.004686 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.020754 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.035492 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.047268 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.061596 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.066416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.066655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.066730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.066801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.066857 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.074149 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.091482 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.103878 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.163162 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.171109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.171315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.171398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.171472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.171551 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.182514 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.183825 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.184067 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.184180 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:30.184142918 +0000 UTC m=+64.769443125 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.201225 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.250550 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.264465 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.273392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.273418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.273426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.273437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.273446 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.287677 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.301838 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.316519 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.335309 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.352795 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.366147 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.376150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.376428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.376652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.376777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.376874 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.486277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.486346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.486370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.486399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.486420 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.522180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.522214 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.522223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.522236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.522244 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.536944 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.540264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.540399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.540479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.540563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.540665 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.551715 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.554881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.554918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.554932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.554950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.554962 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.573700 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.578122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.578156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.578167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.578182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.578191 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.590526 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.593445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.593474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.593482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.593495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.593506 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.605581 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:14Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:14 crc kubenswrapper[4967]: E0120 08:50:14.605789 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.607270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.607307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.607323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.607342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.607357 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.669964 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 11:25:38.435835451 +0000 UTC Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.709990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.710032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.710044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.710061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.710072 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.811801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.811841 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.811853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.811869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.811879 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.915078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.915194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.915219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.915251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:14 crc kubenswrapper[4967]: I0120 08:50:14.915276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:14Z","lastTransitionTime":"2026-01-20T08:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.006914 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/2.log" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.007889 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/1.log" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.014159 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" exitCode=1 Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.014208 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.014239 4967 scope.go:117] "RemoveContainer" containerID="0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.014906 4967 scope.go:117] "RemoveContainer" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.015054 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.018067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.018084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.018092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.018104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.018134 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.029406 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.052256 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.065139 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.073982 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.085505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.099765 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.110335 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.118167 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.120122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.120152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.120168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.120185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.120197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.127141 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.136950 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.149479 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.160687 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.172665 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.186565 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.199439 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.218544 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.224149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.224199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.224216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.224241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.224259 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.232278 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.327187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.327234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.327251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.327270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.327284 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.429671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.429722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.429741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.429766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.429783 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.532680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.532734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.532750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.532771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.532783 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.600621 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.600823 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:50:47.600808263 +0000 UTC m=+82.186108470 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.636383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.636450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.636469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.636496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.636514 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.671188 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 12:23:03.120225791 +0000 UTC Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.693891 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.693982 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.694115 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.694107 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.694327 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.694354 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.694421 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.694511 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.701744 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.701836 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.701880 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.701905 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702000 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702015 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702017 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702039 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702074 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:47.702057045 +0000 UTC m=+82.287357252 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702091 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:47.702083356 +0000 UTC m=+82.287383563 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702085 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702280 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:47.702204669 +0000 UTC m=+82.287504916 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702339 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702360 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702373 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:15 crc kubenswrapper[4967]: E0120 08:50:15.702411 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:50:47.702401515 +0000 UTC m=+82.287701722 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.711402 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.722416 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.735676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.744086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.744119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.744129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.744145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.744156 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.746842 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.760766 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.770722 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.782557 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.792696 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.802544 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.818405 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.836577 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.846336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.846373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.846382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.846406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.846419 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.848079 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.861417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.872488 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.883841 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.896887 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.910898 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:15Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.948221 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.948307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.948330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.948359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:15 crc kubenswrapper[4967]: I0120 08:50:15.948382 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:15Z","lastTransitionTime":"2026-01-20T08:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.018536 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/2.log" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.050710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.050744 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.050753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.050766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.050775 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.153413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.153475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.153533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.153558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.153575 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.256102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.256130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.256138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.256150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.256158 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.359177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.359225 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.359236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.359252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.359264 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.462434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.462491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.462507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.462531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.462548 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.566341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.566402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.566420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.566444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.566462 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.669269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.669338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.669354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.669377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.669394 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.671514 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 19:40:15.025497926 +0000 UTC Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.772408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.772498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.772535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.772565 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.772597 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.877180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.877222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.877233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.877247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.877256 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.979859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.979919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.979943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.979985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:16 crc kubenswrapper[4967]: I0120 08:50:16.980010 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:16Z","lastTransitionTime":"2026-01-20T08:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.081961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.082013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.082026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.082042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.082052 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.185264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.185314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.185328 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.185346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.185358 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.288258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.288317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.288352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.288376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.288394 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.391324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.391409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.391433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.391466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.391490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.494211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.494249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.494258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.494272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.494285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.596921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.596992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.597011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.597038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.597064 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.672325 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 14:47:05.788660326 +0000 UTC Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.693875 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.693943 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.693977 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.693962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:17 crc kubenswrapper[4967]: E0120 08:50:17.694117 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:17 crc kubenswrapper[4967]: E0120 08:50:17.694248 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:17 crc kubenswrapper[4967]: E0120 08:50:17.694327 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:17 crc kubenswrapper[4967]: E0120 08:50:17.694506 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.700017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.700059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.700079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.700107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.700127 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.802686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.802806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.802823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.802853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.802871 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.906391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.906456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.906468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.906483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:17 crc kubenswrapper[4967]: I0120 08:50:17.906494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:17Z","lastTransitionTime":"2026-01-20T08:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.008891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.008923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.008933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.008949 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.008960 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.111660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.111692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.111702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.111716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.111727 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.215302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.215344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.215355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.215375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.215387 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.322835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.322971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.323126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.323172 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.323190 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.426397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.426442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.426453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.426468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.426479 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.528169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.528236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.528258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.528284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.528307 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.632051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.632106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.632115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.632129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.632138 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.673302 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 22:00:18.584165596 +0000 UTC Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.734631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.734704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.734722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.734755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.734772 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.837330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.837578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.837708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.837789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.837858 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.941119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.941153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.941161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.941174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:18 crc kubenswrapper[4967]: I0120 08:50:18.941183 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:18Z","lastTransitionTime":"2026-01-20T08:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.043725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.043782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.043797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.043816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.043830 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.146732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.147140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.147354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.147589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.147860 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.250892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.251211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.251532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.252000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.252215 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.355318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.355354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.355365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.355380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.355389 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.459126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.459429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.459523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.459659 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.459759 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.562808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.563140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.563342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.563558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.563815 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.666683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.666732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.666743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.666763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.666774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.675005 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 14:21:39.710339807 +0000 UTC Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.693692 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:19 crc kubenswrapper[4967]: E0120 08:50:19.693908 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.694263 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:19 crc kubenswrapper[4967]: E0120 08:50:19.694361 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.694571 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:19 crc kubenswrapper[4967]: E0120 08:50:19.694696 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.695003 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:19 crc kubenswrapper[4967]: E0120 08:50:19.695182 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.769646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.769708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.769720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.769738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.769750 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.872652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.873029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.873145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.873268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.873372 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.976200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.976671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.976835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.977038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:19 crc kubenswrapper[4967]: I0120 08:50:19.977223 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:19Z","lastTransitionTime":"2026-01-20T08:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.080084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.080421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.080551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.080707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.080848 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.183575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.183942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.184097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.184285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.184459 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.287400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.287733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.287750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.287766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.287787 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.389356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.389389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.389400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.389415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.389425 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.491472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.491796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.491881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.491985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.492056 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.594322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.594358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.594372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.594391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.594403 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.675714 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:58:01.562231607 +0000 UTC Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.697288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.697346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.697362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.697386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.697402 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.800834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.800896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.800912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.800941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.800959 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.903093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.903159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.903176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.903200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:20 crc kubenswrapper[4967]: I0120 08:50:20.903217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:20Z","lastTransitionTime":"2026-01-20T08:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.006710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.006781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.006798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.006822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.006839 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.110036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.110094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.110112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.110134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.110150 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.213426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.213489 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.213506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.213533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.213553 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.315817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.315870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.315882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.315899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.315911 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.418364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.418664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.418751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.418843 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.418952 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.523063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.523414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.523504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.523623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.523718 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.627086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.627139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.627153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.627178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.627190 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.676389 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 17:16:15.248536582 +0000 UTC Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.693791 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.693788 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.694011 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:21 crc kubenswrapper[4967]: E0120 08:50:21.693940 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.693793 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:21 crc kubenswrapper[4967]: E0120 08:50:21.694222 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:21 crc kubenswrapper[4967]: E0120 08:50:21.694289 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:21 crc kubenswrapper[4967]: E0120 08:50:21.694382 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.730702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.730749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.730764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.730785 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.730798 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.833682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.833721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.833733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.833749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.833761 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.935797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.935865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.935878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.935893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:21 crc kubenswrapper[4967]: I0120 08:50:21.935903 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:21Z","lastTransitionTime":"2026-01-20T08:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.039094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.039157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.039180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.039206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.039225 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.142830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.142875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.142886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.142901 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.142911 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.250825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.250863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.250873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.250891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.250903 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.353583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.353781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.353895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.353971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.353990 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.457313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.457354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.457362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.457380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.457390 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.559522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.559569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.559581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.559597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.559633 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.663203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.663372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.663395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.663735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.663805 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.677489 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 23:17:45.59594546 +0000 UTC Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.771205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.771290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.771316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.771345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.771367 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.874684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.874741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.874758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.874781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.874797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.978581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.978695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.978720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.978750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:22 crc kubenswrapper[4967]: I0120 08:50:22.978771 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:22Z","lastTransitionTime":"2026-01-20T08:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.081102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.081149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.081164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.081186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.081204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.184846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.184910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.184931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.184961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.184982 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.288409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.288498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.288538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.288575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.288599 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.391786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.391856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.391879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.391910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.391934 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.494848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.494913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.494931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.494959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.494977 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.598045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.598083 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.598092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.598107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.598117 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.677717 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 22:24:52.527002343 +0000 UTC Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.694193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.694205 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:23 crc kubenswrapper[4967]: E0120 08:50:23.694731 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.694245 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.694247 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:23 crc kubenswrapper[4967]: E0120 08:50:23.694799 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:23 crc kubenswrapper[4967]: E0120 08:50:23.694596 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:23 crc kubenswrapper[4967]: E0120 08:50:23.695042 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.700630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.700663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.700672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.700683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.700692 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.803726 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.803773 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.803784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.803802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.803813 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.905891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.906149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.906218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.906303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:23 crc kubenswrapper[4967]: I0120 08:50:23.906359 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:23Z","lastTransitionTime":"2026-01-20T08:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.008824 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.008909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.008937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.008965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.008984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.112268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.112311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.112322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.112338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.112350 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.214587 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.214681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.214699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.214725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.214743 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.317969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.318050 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.318076 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.318107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.318131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.421157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.421209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.421220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.421238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.421249 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.523436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.523471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.523481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.523497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.523508 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.625573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.625983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.626167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.626329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.626724 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.678932 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 08:50:13.830774404 +0000 UTC Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.729637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.729684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.729695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.729712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.729723 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.771551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.771595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.771623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.771643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.771655 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.790068 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:24Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.793905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.794134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.794250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.795262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.795488 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.812229 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:24Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.815646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.815680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.815691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.815705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.815716 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.829323 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:24Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.832762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.832789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.832797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.832809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.832818 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.848102 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:24Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.851924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.852068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.852145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.852213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.852280 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.864642 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:24Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:24 crc kubenswrapper[4967]: E0120 08:50:24.864940 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.866751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.867052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.867140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.867229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.867311 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.971180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.971244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.971261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.971285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:24 crc kubenswrapper[4967]: I0120 08:50:24.971305 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:24Z","lastTransitionTime":"2026-01-20T08:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.074329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.074763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.074966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.075158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.075435 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.180035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.180088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.180100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.180116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.180169 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.283132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.283194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.283213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.283239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.283256 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.386576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.386718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.386737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.386912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.386963 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.490741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.491193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.491270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.491350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.491430 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.594035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.594095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.594111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.594133 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.594148 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.679928 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 15:41:39.239149311 +0000 UTC Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.693988 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.694059 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.694197 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:25 crc kubenswrapper[4967]: E0120 08:50:25.694181 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:25 crc kubenswrapper[4967]: E0120 08:50:25.694343 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:25 crc kubenswrapper[4967]: E0120 08:50:25.694498 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.694587 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:25 crc kubenswrapper[4967]: E0120 08:50:25.694775 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.697582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.697669 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.697691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.697712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.697731 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.711875 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.728560 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.745365 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.766114 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.785180 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.799643 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.804136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.804199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.804222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.804251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.804312 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.813463 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.828471 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.847334 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.866498 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.883905 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.898959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.908831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.908877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.908895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.908919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.908937 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:25Z","lastTransitionTime":"2026-01-20T08:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.911871 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.939974 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0847e27388f6a869bf0937d2031b7a9cc4e3d52b07941aa1a0f93d8efada7c53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"message\\\":\\\"me:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0120 08:49:56.833353 6365 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0120 08:49:56.833430 6365 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.963390 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.980746 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:25 crc kubenswrapper[4967]: I0120 08:50:25.995019 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:25Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.012103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.012138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.012149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.012163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.012173 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.115556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.115660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.115685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.115716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.115738 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.218572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.218638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.218650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.218666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.218677 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.321251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.321284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.321292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.321305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.321316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.424275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.424316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.424327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.424343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.424354 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.526168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.526228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.526244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.526263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.526276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.629204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.629291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.629310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.629335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.629352 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.681004 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 22:09:50.670917166 +0000 UTC Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.733084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.733188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.733203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.733226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.733242 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.836523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.836577 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.836588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.836621 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.836634 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.940266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.940332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.940356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.940387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:26 crc kubenswrapper[4967]: I0120 08:50:26.940410 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:26Z","lastTransitionTime":"2026-01-20T08:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.043142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.043409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.043494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.043573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.043656 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.147087 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.147459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.147571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.147713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.147821 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.249853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.249886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.249896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.249912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.249924 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.353337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.353404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.353419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.353440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.353456 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.456242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.456282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.456294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.456310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.456325 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.558771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.558826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.558835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.558849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.558857 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.661495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.661523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.661532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.661545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.661553 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.681507 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 02:49:45.460420867 +0000 UTC Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.693916 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:27 crc kubenswrapper[4967]: E0120 08:50:27.694081 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.694382 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:27 crc kubenswrapper[4967]: E0120 08:50:27.694481 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.695276 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:27 crc kubenswrapper[4967]: E0120 08:50:27.695555 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.699141 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:27 crc kubenswrapper[4967]: E0120 08:50:27.699413 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.763949 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.764196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.764276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.764341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.764406 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.868233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.868760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.868992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.869212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.869403 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.972183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.972254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.972268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.972287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:27 crc kubenswrapper[4967]: I0120 08:50:27.972302 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:27Z","lastTransitionTime":"2026-01-20T08:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.074199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.074246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.074286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.074304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.074313 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.177089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.177155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.177173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.177200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.177218 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.280502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.280563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.280583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.280607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.280655 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.383719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.383750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.383758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.383770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.383797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.486767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.487322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.487446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.487548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.487664 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.589683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.589724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.589739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.589752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.589762 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.682580 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 15:59:15.654183597 +0000 UTC Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.692195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.692268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.692301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.692329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.692351 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.695147 4967 scope.go:117] "RemoveContainer" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" Jan 20 08:50:28 crc kubenswrapper[4967]: E0120 08:50:28.695549 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.705712 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.712271 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.727354 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.740899 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.754662 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.767748 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.778146 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794131 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794901 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.794970 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.807311 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.820563 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.841791 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.857141 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.870130 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.884362 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.896525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.896727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.896826 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.897121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.897249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.897472 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.912643 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.923577 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.933941 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:28Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.999651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.999725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.999740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:28 crc kubenswrapper[4967]: I0120 08:50:28.999759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:28.999771 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:28Z","lastTransitionTime":"2026-01-20T08:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.101352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.101389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.101400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.101416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.101425 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.204778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.204822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.204832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.204850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.204863 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.310348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.310385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.310396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.310411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.310424 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.413234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.413269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.413280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.413295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.413307 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.515226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.515261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.515270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.515283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.515293 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.617414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.617449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.617458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.617474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.617485 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.683103 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 21:14:33.756223351 +0000 UTC Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.693653 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.693704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.693759 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.693654 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:29 crc kubenswrapper[4967]: E0120 08:50:29.693780 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:29 crc kubenswrapper[4967]: E0120 08:50:29.693866 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:29 crc kubenswrapper[4967]: E0120 08:50:29.693936 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:29 crc kubenswrapper[4967]: E0120 08:50:29.694019 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.719179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.719218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.719232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.719249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.719261 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.822810 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.822882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.822907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.822936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.822959 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.925307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.925363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.925377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.925398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:29 crc kubenswrapper[4967]: I0120 08:50:29.925409 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:29Z","lastTransitionTime":"2026-01-20T08:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.028341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.028416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.028439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.028470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.028492 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.131274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.131324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.131339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.131362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.131386 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.234015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.234053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.234064 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.234080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.234091 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.268672 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:30 crc kubenswrapper[4967]: E0120 08:50:30.268882 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:30 crc kubenswrapper[4967]: E0120 08:50:30.269016 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:51:02.268984405 +0000 UTC m=+96.854284652 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.336391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.336447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.336465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.336492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.336513 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.438896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.438932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.438939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.438953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.438963 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.542128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.542175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.542187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.542203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.542215 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.644880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.644933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.644948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.644965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.644979 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.676477 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.677254 4967 scope.go:117] "RemoveContainer" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" Jan 20 08:50:30 crc kubenswrapper[4967]: E0120 08:50:30.677418 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.684156 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 19:24:00.295766789 +0000 UTC Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.747046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.747088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.747100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.747116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.747128 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.849345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.849402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.849423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.849451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.849475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.952706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.952742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.952750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.952767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:30 crc kubenswrapper[4967]: I0120 08:50:30.952776 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:30Z","lastTransitionTime":"2026-01-20T08:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.055178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.055237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.055253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.055275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.055292 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.158294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.158369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.158378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.158393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.158428 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.260715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.260762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.260772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.260789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.260803 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.362827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.362870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.362882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.362900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.362912 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.465410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.465462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.465477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.465499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.465515 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.568288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.568327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.568338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.568353 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.568363 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.670667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.670713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.670742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.670760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.670772 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.684330 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 09:44:52.107731315 +0000 UTC Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.693660 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.693688 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.693781 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.693804 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:31 crc kubenswrapper[4967]: E0120 08:50:31.693907 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:31 crc kubenswrapper[4967]: E0120 08:50:31.694029 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:31 crc kubenswrapper[4967]: E0120 08:50:31.694160 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:31 crc kubenswrapper[4967]: E0120 08:50:31.694278 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.773448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.773495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.773507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.773525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.773539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.875898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.875934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.875945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.875959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.875971 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.978501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.978563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.978575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.978591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:31 crc kubenswrapper[4967]: I0120 08:50:31.978602 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:31Z","lastTransitionTime":"2026-01-20T08:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.075372 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/0.log" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.075418 4967 generic.go:334] "Generic (PLEG): container finished" podID="7ba8630f-92bc-4708-a722-a7e27c747073" containerID="650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d" exitCode=1 Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.075443 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerDied","Data":"650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.075796 4967 scope.go:117] "RemoveContainer" containerID="650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.080220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.080260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.080278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.080300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.080316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.084934 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.095489 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.105892 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.124013 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.139218 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.164280 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.177676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.182712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.182760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.182770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.182789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.182806 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.188927 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.201053 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.212629 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.225349 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.235227 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.245676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.256586 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.269368 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.280328 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.284808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.284940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.285005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.285079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.285159 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.291847 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.302758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:32Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.387913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.387965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.387974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.387990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.387999 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.489768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.490109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.490290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.490418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.490552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.596936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.596980 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.596990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.597013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.597025 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.685054 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 21:14:07.13794221 +0000 UTC Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.704521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.704729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.704833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.705028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.705123 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.893188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.893597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.893890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.894067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:32 crc kubenswrapper[4967]: I0120 08:50:32.894223 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:32Z","lastTransitionTime":"2026-01-20T08:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.005024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.005073 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.005085 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.005103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.005116 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.104024 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/0.log" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.104077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerStarted","Data":"124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.110366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.110432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.110456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.110486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.110510 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.123161 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.144173 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.179455 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.212999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.213250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.213348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.213445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.213519 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.214556 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.236699 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.257106 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.270524 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.280127 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.290748 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.302503 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.314971 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.315990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.316027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.316036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.316055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.316066 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.328177 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.340410 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.351693 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.363944 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.373509 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.386879 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.399600 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:33Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.418089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.418124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.418133 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.418146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.418156 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.520432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.520495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.520514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.520541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.520560 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.622781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.622822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.622833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.622848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.622859 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.685629 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 11:36:11.2987535 +0000 UTC Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.693414 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.693482 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:33 crc kubenswrapper[4967]: E0120 08:50:33.693579 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.693442 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:33 crc kubenswrapper[4967]: E0120 08:50:33.693681 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:33 crc kubenswrapper[4967]: E0120 08:50:33.693740 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.693994 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:33 crc kubenswrapper[4967]: E0120 08:50:33.694116 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.725081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.725327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.725440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.725516 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.725577 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.827891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.828287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.828575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.828900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.829193 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.931939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.931986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.931996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.932011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:33 crc kubenswrapper[4967]: I0120 08:50:33.932020 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:33Z","lastTransitionTime":"2026-01-20T08:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.034731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.035139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.035465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.036168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.036224 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.139241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.139684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.139923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.140146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.140502 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.242764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.242859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.242878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.242914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.242927 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.345384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.345435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.345448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.345469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.345483 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.448159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.448203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.448213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.448229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.448240 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.550635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.550700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.550719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.550745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.550763 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.653516 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.653657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.653674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.653692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.653704 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.687156 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 08:00:20.31443843 +0000 UTC Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.755870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.755915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.755926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.755943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.755955 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.858330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.858380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.858395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.858416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.858431 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.960191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.960233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.960244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.960259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:34 crc kubenswrapper[4967]: I0120 08:50:34.960274 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:34Z","lastTransitionTime":"2026-01-20T08:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.062941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.062975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.062986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.063005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.063015 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.164890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.164923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.164932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.164970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.164979 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.202079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.202109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.202117 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.202149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.202158 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.213921 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.217766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.217806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.217815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.217828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.217837 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.228595 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.232092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.232157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.232171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.232184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.232197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.247732 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.250955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.250979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.250987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.251001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.251009 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.264050 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.267204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.267237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.267250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.267289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.267302 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.282672 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.282790 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.283782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.283811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.283836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.283850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.283858 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.385764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.385825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.385835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.385849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.385857 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.488218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.488277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.488294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.488319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.488336 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.590410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.590444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.590452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.590467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.590475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.688250 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 07:04:00.140576411 +0000 UTC Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693125 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693135 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693168 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.693257 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693317 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.693391 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.693332 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.693489 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:35 crc kubenswrapper[4967]: E0120 08:50:35.693519 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.707262 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.721008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.733315 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.743732 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.754962 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.768670 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.785382 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.795223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.795256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.795265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.795278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.795287 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.800651 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.811704 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.823884 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.833558 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.843822 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.854350 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.864885 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.875435 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.882984 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.893011 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.897456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.897496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.897506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.897520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.897531 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.903873 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:35Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.999277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.999314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.999324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.999338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:35 crc kubenswrapper[4967]: I0120 08:50:35.999348 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:35Z","lastTransitionTime":"2026-01-20T08:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.101695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.101755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.101775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.101798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.101817 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.203884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.203992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.204013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.204038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.204054 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.306269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.306308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.306318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.306332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.306344 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.407981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.408027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.408059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.408079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.408092 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.510866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.510904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.510913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.510928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.510939 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.613518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.613564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.613575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.613592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.613605 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.688878 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 14:31:36.355596362 +0000 UTC Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.716485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.716517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.716526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.716536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.716546 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.818798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.818825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.818835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.818847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.818893 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.920830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.920912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.920935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.920971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:36 crc kubenswrapper[4967]: I0120 08:50:36.920995 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:36Z","lastTransitionTime":"2026-01-20T08:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.023848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.023895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.023904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.023917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.023926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.126747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.126787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.126798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.126812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.126822 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.231048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.231090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.231100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.231307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.231317 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.334129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.334170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.334181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.334196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.334207 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.437308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.437352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.437362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.437379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.437390 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.540966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.541058 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.541075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.541361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.541382 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.645334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.645381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.645429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.645448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.645463 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.689143 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 18:07:54.240708723 +0000 UTC Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.693538 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.693587 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.693673 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.693729 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:37 crc kubenswrapper[4967]: E0120 08:50:37.693799 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:37 crc kubenswrapper[4967]: E0120 08:50:37.693966 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:37 crc kubenswrapper[4967]: E0120 08:50:37.694030 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:37 crc kubenswrapper[4967]: E0120 08:50:37.694140 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.747183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.747238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.747249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.747266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.747279 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.850149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.850191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.850203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.850242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.850254 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.952015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.952075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.952086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.952102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:37 crc kubenswrapper[4967]: I0120 08:50:37.952113 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:37Z","lastTransitionTime":"2026-01-20T08:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.054410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.054471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.054488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.054510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.054525 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.156922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.157017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.157040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.157072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.157094 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.260090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.260141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.260154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.260174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.260186 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.362881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.362925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.362956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.362973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.362984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.464990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.465036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.465045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.465060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.465070 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.567909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.567978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.568002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.568033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.568059 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.670716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.670762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.670786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.670812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.670827 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.690067 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 12:39:57.90426249 +0000 UTC Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.774024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.774065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.774074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.774093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.774102 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.877183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.877231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.877243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.877262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.877273 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.978937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.978973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.978981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.978995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:38 crc kubenswrapper[4967]: I0120 08:50:38.979004 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:38Z","lastTransitionTime":"2026-01-20T08:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.081568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.081651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.081673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.081700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.081722 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.184291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.184326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.184336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.184352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.184362 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.286849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.286907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.286923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.286945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.286961 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.389379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.389427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.389439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.389453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.389465 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.491747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.491805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.491820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.491844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.491870 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.594085 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.594171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.594196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.594217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.594231 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.690444 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 00:30:30.430150958 +0000 UTC Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.693898 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.693919 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.693894 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:39 crc kubenswrapper[4967]: E0120 08:50:39.694024 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:39 crc kubenswrapper[4967]: E0120 08:50:39.694098 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.694145 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:39 crc kubenswrapper[4967]: E0120 08:50:39.694263 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:39 crc kubenswrapper[4967]: E0120 08:50:39.694329 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.696483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.696545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.696557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.696572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.696585 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.802640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.802731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.802747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.802807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.802825 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.906239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.906291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.906307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.906365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:39 crc kubenswrapper[4967]: I0120 08:50:39.906383 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:39Z","lastTransitionTime":"2026-01-20T08:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.009331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.009393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.009417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.009445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.009466 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.112258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.112303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.112324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.112345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.112361 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.215251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.215286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.215297 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.215313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.215324 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.319119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.319163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.319179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.319201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.319217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.421604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.421656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.421666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.421682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.421694 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.525377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.525447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.525470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.525501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.525524 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.628450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.628500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.628518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.628542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.628560 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.690999 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 09:37:44.977295431 +0000 UTC Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.731657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.731691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.731720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.731741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.731755 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.834600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.834692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.834709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.834732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.834751 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.937425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.937483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.937498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.937521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:40 crc kubenswrapper[4967]: I0120 08:50:40.937535 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:40Z","lastTransitionTime":"2026-01-20T08:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.039876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.039915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.039923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.039938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.039947 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.142425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.142459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.142468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.142482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.142491 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.245045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.245099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.245111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.245130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.245144 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.348100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.348144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.348154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.348169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.348180 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.451170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.451212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.451266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.451290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.451302 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.553568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.553630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.553642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.553660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.553672 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.656396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.656437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.656448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.656465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.656476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.691533 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 14:03:52.614098574 +0000 UTC Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.693916 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.694017 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:41 crc kubenswrapper[4967]: E0120 08:50:41.694050 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.694107 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.694108 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:41 crc kubenswrapper[4967]: E0120 08:50:41.694221 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:41 crc kubenswrapper[4967]: E0120 08:50:41.694467 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:41 crc kubenswrapper[4967]: E0120 08:50:41.695057 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.695500 4967 scope.go:117] "RemoveContainer" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.759346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.759381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.759393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.759411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.759422 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.862005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.862071 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.862093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.862123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.862144 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.965135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.965201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.965223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.965250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:41 crc kubenswrapper[4967]: I0120 08:50:41.965271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:41Z","lastTransitionTime":"2026-01-20T08:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.067789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.067829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.067841 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.067860 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.067874 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.141689 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/2.log" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.145291 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.146648 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.167986 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.170697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.170729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.170743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.170762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.170775 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.184876 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.198018 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.210490 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.224043 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.233785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.248222 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.264216 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.272991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.273025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.273034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.273050 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.273060 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.278068 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.290298 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.304478 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.319134 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.342240 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.364435 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.374883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.374924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.374935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.374953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.374964 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.378463 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.391061 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.403116 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.413735 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:42Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.477793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.477831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.477840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.477856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.477866 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.580738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.580770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.580778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.580791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.580800 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.682734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.682756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.682765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.682778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.682788 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.691606 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 11:01:28.83430202 +0000 UTC Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.784878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.784909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.784917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.784933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.784942 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.887001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.887035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.887045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.887059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.887068 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.989545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.989651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.989679 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.989708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:42 crc kubenswrapper[4967]: I0120 08:50:42.989727 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:42Z","lastTransitionTime":"2026-01-20T08:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.092725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.092791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.092804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.092819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.092830 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.150052 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/3.log" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.150780 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/2.log" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.154012 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" exitCode=1 Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.154062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.154119 4967 scope.go:117] "RemoveContainer" containerID="61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.154925 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:50:43 crc kubenswrapper[4967]: E0120 08:50:43.155260 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.171794 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.182969 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.195402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.195470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.195482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.195502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.195516 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.197260 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.213072 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.224300 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.236106 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.248921 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.258282 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.268674 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.288316 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61812fda5be21e43c71e88866fe3c7ff0e55e9a2be657090f2489d4fa7407721\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:14Z\\\",\\\"message\\\":\\\"68 6584 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0120 08:50:14.568378 6584 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.568531 6584 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0120 08:50:14.571736 6584 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0120 08:50:14.571794 6584 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0120 08:50:14.571823 6584 handler.go:208] Removed *v1.Node event handler 2\\\\nI0120 08:50:14.571859 6584 factory.go:656] Stopping watch factory\\\\nI0120 08:50:14.571870 6584 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0120 08:50:14.621731 6584 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0120 08:50:14.621766 6584 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0120 08:50:14.621850 6584 ovnkube.go:599] Stopped ovnkube\\\\nI0120 08:50:14.621881 6584 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0120 08:50:14.621965 6584 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:42Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0120 08:50:42.710156 6989 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0120 08:50:42.710164 6989 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 590.856µs\\\\nI0120 08:50:42.710191 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-h8vz7\\\\nI0120 08:50:42.710204 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710211 6989 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710217 6989 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0120 08:50:42.710232 6989 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF0120 08:50:42.710247 6989 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.297478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.297741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.297808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.297883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.297958 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.302372 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.313250 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.323531 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.332069 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.341971 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.350966 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.360226 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.369399 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:43Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.400520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.400578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.400588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.400603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.400634 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.502984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.503065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.503105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.503143 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.503166 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.605500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.605539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.605549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.605566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.605577 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.692083 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:28:01.127840182 +0000 UTC Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.693364 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.693399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:43 crc kubenswrapper[4967]: E0120 08:50:43.693506 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.693555 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.693560 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:43 crc kubenswrapper[4967]: E0120 08:50:43.693675 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:43 crc kubenswrapper[4967]: E0120 08:50:43.693849 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:43 crc kubenswrapper[4967]: E0120 08:50:43.693945 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.708079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.708115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.708126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.708142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.708155 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.711604 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.811068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.811333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.811405 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.811499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.811595 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.914965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.915040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.915063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.915094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:43 crc kubenswrapper[4967]: I0120 08:50:43.915120 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:43Z","lastTransitionTime":"2026-01-20T08:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.018127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.018162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.018173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.018189 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.018200 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.120825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.121257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.121781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.122037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.122271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.160060 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/3.log" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.163845 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:50:44 crc kubenswrapper[4967]: E0120 08:50:44.164006 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.184216 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be2a3a2f-f5a5-418b-8752-d1f017f0e2f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5879f9d52ca4ec745e9adc618ca74f9c1ecd395bcfc85b01adc06e5ae93a6da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb232f2ebcd144a969f5fc3cc5b192a4454434917f0a7700d9f1373cece0253\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efae4da5f2d2a14a06c78e20cc23d2811dbecf7a66eb86ee731a74e2c88ff4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09e324d3f818853cf74667a9c020ce195feb13f3535d6c29973c681e3553a3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cc6b7e0c06a081516d203051e26fe1cb4a5febf9d273aaea117b284fe423c69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.195948 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.208660 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.220174 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.225467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.225490 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.225500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.225515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.225525 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.233112 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.243542 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.256653 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.271306 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.288041 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.298825 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.310797 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.326038 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.327579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.327662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.327688 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.327713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.327737 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.340510 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.355670 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.366582 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.377005 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.393901 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:42Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0120 08:50:42.710156 6989 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0120 08:50:42.710164 6989 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 590.856µs\\\\nI0120 08:50:42.710191 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-h8vz7\\\\nI0120 08:50:42.710204 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710211 6989 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710217 6989 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0120 08:50:42.710232 6989 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF0120 08:50:42.710247 6989 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.423012 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.430027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.430065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.430078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.430094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.430104 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.433522 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:44Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.532014 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.532074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.532090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.532113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.532132 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.635200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.635253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.635262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.635278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.635287 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.692909 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 18:57:05.644489221 +0000 UTC Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.736883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.736928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.736938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.736952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.736963 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.840087 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.840170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.840224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.840249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.840262 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.942446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.942483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.942491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.942507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:44 crc kubenswrapper[4967]: I0120 08:50:44.942517 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:44Z","lastTransitionTime":"2026-01-20T08:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.045602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.045672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.045683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.045703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.045725 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.148848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.148925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.148948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.148979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.148997 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.252022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.252064 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.252076 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.252091 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.252101 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.354506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.354571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.354593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.354660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.354685 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.457965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.457997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.458008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.458024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.458034 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.521129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.521451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.521465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.521482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.521493 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.538227 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.542502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.542588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.542676 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.542694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.542705 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.559214 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.562916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.563028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.563049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.563073 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.563089 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.581372 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.585977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.586012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.586021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.586035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.586044 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.596858 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.599985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.600025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.600034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.600047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.600056 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.613013 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.613228 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.614793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.614838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.614853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.614879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.614895 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.693206 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 02:03:27.311668934 +0000 UTC Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.693372 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.693398 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.693416 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.693487 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.693552 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.693664 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.693720 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:45 crc kubenswrapper[4967]: E0120 08:50:45.693678 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.705135 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.714887 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.717063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.717090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.717100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.717114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.717123 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.726585 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.741146 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.755238 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.767416 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.784412 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.794040 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.803084 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822236 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.822690 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:42Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0120 08:50:42.710156 6989 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0120 08:50:42.710164 6989 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 590.856µs\\\\nI0120 08:50:42.710191 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-h8vz7\\\\nI0120 08:50:42.710204 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710211 6989 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710217 6989 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0120 08:50:42.710232 6989 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF0120 08:50:42.710247 6989 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.835171 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.845146 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.855792 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.865078 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.875717 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.886394 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.897179 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.914570 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be2a3a2f-f5a5-418b-8752-d1f017f0e2f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5879f9d52ca4ec745e9adc618ca74f9c1ecd395bcfc85b01adc06e5ae93a6da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb232f2ebcd144a969f5fc3cc5b192a4454434917f0a7700d9f1373cece0253\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efae4da5f2d2a14a06c78e20cc23d2811dbecf7a66eb86ee731a74e2c88ff4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09e324d3f818853cf74667a9c020ce195feb13f3535d6c29973c681e3553a3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cc6b7e0c06a081516d203051e26fe1cb4a5febf9d273aaea117b284fe423c69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.925325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.925373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.925384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.925401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.925411 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:45Z","lastTransitionTime":"2026-01-20T08:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:45 crc kubenswrapper[4967]: I0120 08:50:45.928970 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:45Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.027835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.027872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.027880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.027893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.027901 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.131437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.131570 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.131591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.131660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.131683 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.234750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.234789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.234803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.234820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.234832 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.337207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.337254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.337268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.337287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.337302 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.439650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.439705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.439719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.439738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.439752 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.541823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.541864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.541878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.541921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.541936 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.644426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.644500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.644521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.644551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.644573 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.694232 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 10:56:09.436464442 +0000 UTC Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.746513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.746608 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.746675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.746707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.746730 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.849882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.849922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.849931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.849945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.849955 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.952659 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.952728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.952749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.952780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:46 crc kubenswrapper[4967]: I0120 08:50:46.952802 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:46Z","lastTransitionTime":"2026-01-20T08:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.055323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.055380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.055398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.055420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.055437 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.158810 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.158909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.158926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.158950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.158967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.265988 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.266028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.266040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.266056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.266066 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.368502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.368538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.368547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.368561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.368570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.470973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.471034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.471044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.471057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.471066 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.573346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.573395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.573405 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.573420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.573430 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.653280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.653401 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.653378294 +0000 UTC m=+146.238678501 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.675863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.675910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.675924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.675942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.675952 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.693117 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.693173 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.693253 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.693277 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.693303 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.693429 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.693488 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.693700 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.694736 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 08:28:49.066253686 +0000 UTC Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.754262 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.754354 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.754401 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.754450 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.754979 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.754999 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755030 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755077 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755016 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755076 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755153 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.755124374 +0000 UTC m=+146.340424621 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755163 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755208 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755238 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.755215366 +0000 UTC m=+146.340515613 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755270 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.755254307 +0000 UTC m=+146.340554614 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 20 08:50:47 crc kubenswrapper[4967]: E0120 08:50:47.755299 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.755285548 +0000 UTC m=+146.340585795 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.778230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.778263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.778273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.778290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.778300 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.880688 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.880739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.880750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.880766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.880777 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.983340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.983383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.983395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.983412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:47 crc kubenswrapper[4967]: I0120 08:50:47.983425 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:47Z","lastTransitionTime":"2026-01-20T08:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.085561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.085628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.085640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.085660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.085671 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.189193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.189233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.189248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.189264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.189275 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.291928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.292215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.292224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.292236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.292260 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.394996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.395032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.395040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.395054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.395063 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.497453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.497488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.497501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.497518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.497531 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.600498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.600538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.600547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.600561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.600570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.694982 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 06:40:33.159484801 +0000 UTC Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.703184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.703230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.703247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.703269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.703286 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.805847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.805919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.805941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.805966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.805983 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.908460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.908543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.908558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.908576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:48 crc kubenswrapper[4967]: I0120 08:50:48.908637 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:48Z","lastTransitionTime":"2026-01-20T08:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.011193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.011229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.011243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.011259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.011268 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.114438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.114482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.114491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.114507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.114519 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.217505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.217564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.217580 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.217602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.217659 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.320521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.320578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.320597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.320654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.320676 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.424132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.424210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.424223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.424238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.424248 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.527451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.527510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.527528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.527552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.527575 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.630544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.630603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.630691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.630723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.630759 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.693536 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.693669 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.693573 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:49 crc kubenswrapper[4967]: E0120 08:50:49.693784 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.693847 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:49 crc kubenswrapper[4967]: E0120 08:50:49.693956 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:49 crc kubenswrapper[4967]: E0120 08:50:49.694055 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:49 crc kubenswrapper[4967]: E0120 08:50:49.694248 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.695632 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 19:57:50.633893069 +0000 UTC Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.734372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.734448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.734470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.734499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.734520 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.838407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.838497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.838522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.838552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.838574 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.941366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.941433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.941456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.941489 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:49 crc kubenswrapper[4967]: I0120 08:50:49.941513 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:49Z","lastTransitionTime":"2026-01-20T08:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.045240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.045281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.045326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.045362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.045380 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.148931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.149011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.149028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.149054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.149073 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.251874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.251945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.251969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.251997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.252018 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.355270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.355322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.355353 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.355396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.355424 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.459504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.459584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.459601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.459657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.459674 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.561998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.562079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.562101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.562131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.562154 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.665555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.665605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.665651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.665667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.665678 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.696748 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 22:57:45.015156454 +0000 UTC Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.768497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.768572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.768586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.768606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.768646 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.871207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.871318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.871356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.871395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.871418 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.974563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.974683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.974720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.974752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:50 crc kubenswrapper[4967]: I0120 08:50:50.974774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:50Z","lastTransitionTime":"2026-01-20T08:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.077682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.077763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.077785 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.077821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.077847 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.181515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.181584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.181595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.181638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.181650 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.284069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.284116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.284131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.284153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.284168 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.387757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.388074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.388321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.388530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.388774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.492013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.492089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.492113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.492141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.492159 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.595525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.595568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.595579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.595597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.595627 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.693879 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:51 crc kubenswrapper[4967]: E0120 08:50:51.694084 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.694164 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.694210 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.694210 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:51 crc kubenswrapper[4967]: E0120 08:50:51.694480 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:51 crc kubenswrapper[4967]: E0120 08:50:51.694580 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:51 crc kubenswrapper[4967]: E0120 08:50:51.694679 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.697142 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 08:32:06.379258323 +0000 UTC Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.698074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.698114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.698130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.698151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.698168 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.802196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.802262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.802283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.802318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.802361 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.905100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.905486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.906059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.906256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:51 crc kubenswrapper[4967]: I0120 08:50:51.906405 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:51Z","lastTransitionTime":"2026-01-20T08:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.009908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.009946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.009956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.009971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.009981 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.113088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.113144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.113162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.113187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.113204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.215604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.215693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.215712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.215738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.215755 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.318276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.318349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.318371 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.318404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.318426 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.421362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.421446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.421479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.421508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.421528 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.525211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.525270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.525292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.525321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.525341 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.628295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.628367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.628390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.628417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.628439 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.697812 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 07:43:28.645645307 +0000 UTC Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.731415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.731445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.731456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.731474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.731487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.833533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.833568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.833579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.833596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.833621 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.936748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.936795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.936807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.936824 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:52 crc kubenswrapper[4967]: I0120 08:50:52.936837 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:52Z","lastTransitionTime":"2026-01-20T08:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.039748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.039793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.039807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.039827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.039843 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.142650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.142701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.142712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.142730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.142741 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.245368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.245396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.245404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.245418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.245426 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.348562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.348647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.348665 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.348697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.348735 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.451707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.451744 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.451752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.451766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.451793 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.554682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.554755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.554780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.554809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.554833 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.657310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.657592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.657814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.657964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.658088 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.693681 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.693760 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:53 crc kubenswrapper[4967]: E0120 08:50:53.693888 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.693956 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.693970 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:53 crc kubenswrapper[4967]: E0120 08:50:53.694161 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:53 crc kubenswrapper[4967]: E0120 08:50:53.694211 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:53 crc kubenswrapper[4967]: E0120 08:50:53.694315 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.698322 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 23:51:04.277700314 +0000 UTC Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.760470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.760535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.760554 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.760575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.760592 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.863837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.863902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.863925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.863954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.863977 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.967166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.967229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.967252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.967279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:53 crc kubenswrapper[4967]: I0120 08:50:53.967299 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:53Z","lastTransitionTime":"2026-01-20T08:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.069191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.069261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.069285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.069357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.069383 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.171907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.171995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.172028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.172057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.172078 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.274697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.274765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.274784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.274808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.274828 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.377905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.377949 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.377959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.377974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.377986 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.481494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.481556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.481574 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.481602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.481657 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.585293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.585376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.585399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.585432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.585455 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.688080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.688130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.688145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.688162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.688174 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.699555 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 05:22:51.533323887 +0000 UTC Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.791404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.791451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.791463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.791506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.791524 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.895077 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.895145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.895167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.895196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.895218 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.998739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.998808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.998830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.998858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:54 crc kubenswrapper[4967]: I0120 08:50:54.998884 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:54Z","lastTransitionTime":"2026-01-20T08:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.100461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.100523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.100533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.100548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.100558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.202647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.202689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.202702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.202722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.203005 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.305707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.305746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.305756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.305794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.305804 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.408497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.408764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.408832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.408910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.409002 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.511018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.511597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.511840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.511989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.512131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.614063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.614101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.614112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.614128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.614139 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.693267 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.693744 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.693892 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.693961 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.694001 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.694103 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.694335 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.694418 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.699730 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 13:19:52.389200198 +0000 UTC Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.706796 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f3f2ed7981a449f05585e8f2e2334408fb6db985caa68a3209f75569c535cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716312 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8vz7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fbe33-7bc2-451f-9c86-bbf2bb8f37b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11d3ddf8b53a23ba16f27277a93b4c049c7aefb88c8db97880ffae645e77a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ltlch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:48Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8vz7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.716752 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.732775 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf90c10e-f854-47e6-bd6b-97932e312bec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://252b039d37b38701ab86d37d7d7b77fcf131bf1dd6da3e8b099e7a7cee7a7ee5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4f3223fbd1126119350468d38f4e4e0f3428f84282e704294df351af5f658c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.741947 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq2xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lxp6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.753784 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e03e4415-da9f-4306-8194-30d9bbc66a86\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"le observer\\\\nW0120 08:49:43.435749 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0120 08:49:43.435986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0120 08:49:43.447460 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3405979623/tls.crt::/tmp/serving-cert-3405979623/tls.key\\\\\\\"\\\\nI0120 08:49:43.670364 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0120 08:49:43.672750 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0120 08:49:43.672771 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0120 08:49:43.672792 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0120 08:49:43.672798 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0120 08:49:43.678183 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0120 08:49:43.678207 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678213 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0120 08:49:43.678219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0120 08:49:43.678223 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0120 08:49:43.678226 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0120 08:49:43.678231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0120 08:49:43.678230 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0120 08:49:43.680490 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.764703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.764869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.764886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.764911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.764928 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.766123 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.778373 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.778911 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.781984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.782063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.782089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.782119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.782145 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.795821 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8958ac5a980d93aa06fc23c7c31a91faf4f646b21eaf9f55dbf5a19f7f734b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9110796a758655530e344c3ec3ff2e370f99d924f678629d3723281daf8cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.797346 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.803115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.803291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.803355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.803426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.803487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.805633 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nxlff" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73537611-ae3a-4699-a02f-da4f5192d702\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d027b8193b252adad46c53c775c734133f437067a491101cba0ac38fdfd85bbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22fng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nxlff\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.814400 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.815410 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5de94293-c4d1-4169-a8df-69843d57278a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db03a0aa6eaee0fd36a32591bc18bcf9e9affdc9920b1dade6e9ec15bcf71f3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ttlgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bcgrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.817875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.817903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.817914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.817929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.817940 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.832197 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.832724 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9f22bb-157c-4556-9066-cd3cdacdf3af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:42Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0120 08:50:42.710156 6989 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0120 08:50:42.710164 6989 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 590.856µs\\\\nI0120 08:50:42.710191 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-h8vz7\\\\nI0120 08:50:42.710204 6989 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710211 6989 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0120 08:50:42.710217 6989 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0120 08:50:42.710232 6989 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF0120 08:50:42.710247 6989 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:50:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49rck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fc9bg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.836367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.836406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.836422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.836445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.836461 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.845514 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb9067c-29d9-4392-8c7e-5198eb5ee93d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc579d5cf0bacf12f6fdf3bc344e744e2ddebbbec684f1d7e88593c88fa54345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18cac8dd3ec63db621d7e35f5dd6e06b656a93b45b4a9215d412a222771b52ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2ad796523ec13fd6e7dbae5d6763eee136ca1c0894b7d2d13f326b6af0a92e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f34b8bc9bd9918bfc20376ffc9819f38fe677062bc2b2bae12c6d47a897ad73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98448be2172273267333fc0263cc5f564620a6d9cad7d6e9da2ef9339f6fdf6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://299ab52d8adf5920ba057db16010a48c3d4ee13b87d2b08cd1021cc1b9f2e9ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670ab213d474c930abd9ff2070ffa8b80bc02fc6b7e9d28410048eaa0f174b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gr52w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hlcgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.849673 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4107ff2b-f75c-4198-af98-8ba98c029f9d\\\",\\\"systemUUID\\\":\\\"127fa9ec-49e3-48ce-8fe1-c4315b50082f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: E0120 08:50:55.849787 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.851156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.851177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.851185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.851197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.851206 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.857757 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b70968fe-6e57-4538-a1e3-eec35e058cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57da8a5579e3af41fd1479c7d583e4dbbb797cbfd55647b1c11e23d7d240720e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f74aa57e2881d7f7d74d43dfadf7df2a0b24e4871100bbcebce181ec99b11a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f7d0d81b3583a3579c5c34e5a67cdc3bd183b8c41da62616baad35faeb67ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d808bdce335b96f9008cfd0e7bde8b14630ff54a7b367f16c9e4a42a639e79b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.867685 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5c4e15f-99b5-46ed-93bf-15b3413e6eab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25f73c2cf99a22c8d2de096ec66d75b1bd9a9f84657d21e3c63313f06ff39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6cda6a07fe70875bac03c62315480ea7401f5531c666215fec51d993855d538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xhd48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mfb5x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.878705 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1f8e127-e377-476d-9281-5b1dca7b2e66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63b322b14cf7e49cf010fef92bc2186e74b50fcc828673573d6213631da665c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b194825ebc0289f19f8af15f79821ecceaa329ea0b87bd72633f3f6119b6e8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9058966e6d08f2e7abd0815be7e2066cba2450f6f565d7f74972a7c8ae6f1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.891404 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc02ae093e5605cb56038e700436999477951cdc5c8e31bba6f7ac680f482da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.907989 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.923327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6wm4t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ba8630f-92bc-4708-a722-a7e27c747073\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:50:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-20T08:50:31Z\\\",\\\"message\\\":\\\"2026-01-20T08:49:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd\\\\n2026-01-20T08:49:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e4e55371-d7ad-473c-a210-c182f589f4dd to /host/opt/cni/bin/\\\\n2026-01-20T08:49:46Z [verbose] multus-daemon started\\\\n2026-01-20T08:49:46Z [verbose] Readiness Indicator file check\\\\n2026-01-20T08:50:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:50:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgmkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6wm4t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.943778 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be2a3a2f-f5a5-418b-8752-d1f017f0e2f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-20T08:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5879f9d52ca4ec745e9adc618ca74f9c1ecd395bcfc85b01adc06e5ae93a6da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4eb232f2ebcd144a969f5fc3cc5b192a4454434917f0a7700d9f1373cece0253\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efae4da5f2d2a14a06c78e20cc23d2811dbecf7a66eb86ee731a74e2c88ff4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09e324d3f818853cf74667a9c020ce195feb13f3535d6c29973c681e3553a3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cc6b7e0c06a081516d203051e26fe1cb4a5febf9d273aaea117b284fe423c69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-20T08:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ff77d663183d02cbb09871a116e93b45ec044acfeae3bd004d0477d23a2dfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfeb98da8c864e9741b068cd3eb10583e0e42842e9f4cba441c0d299fe05953c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ea97793ffc507d4de6cebaf942a5f0c1c83583c35ebbc25917d854e7dd43189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-20T08:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-20T08:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-20T08:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-20T08:50:55Z is after 2025-08-24T17:21:41Z" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.953635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.953673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.953685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.953704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:55 crc kubenswrapper[4967]: I0120 08:50:55.953715 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:55Z","lastTransitionTime":"2026-01-20T08:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.057094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.057138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.057155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.057175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.057192 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.158747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.158786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.158795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.158811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.158820 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.261940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.262100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.262113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.262131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.262142 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.364667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.364720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.364736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.364758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.364773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.466569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.466648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.466671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.466692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.466714 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.568759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.568821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.568838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.568861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.568878 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.671190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.671248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.671266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.671290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.671316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.700188 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 23:14:19.646199446 +0000 UTC Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.774226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.774270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.774282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.774298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.774308 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.877242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.877294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.877309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.877331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.877347 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.979789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.979829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.979837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.979854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:56 crc kubenswrapper[4967]: I0120 08:50:56.979868 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:56Z","lastTransitionTime":"2026-01-20T08:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.082410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.082465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.082485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.082534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.082557 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.184820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.184865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.184875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.184889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.184902 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.293396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.293423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.293431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.293444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.293453 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.395646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.395672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.395680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.395694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.395703 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.500176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.500238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.500255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.500279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.500295 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.603228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.603282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.603304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.603332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.603356 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.694125 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.694223 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.694156 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:57 crc kubenswrapper[4967]: E0120 08:50:57.694358 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:57 crc kubenswrapper[4967]: E0120 08:50:57.694522 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.694605 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:57 crc kubenswrapper[4967]: E0120 08:50:57.694802 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.695452 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:50:57 crc kubenswrapper[4967]: E0120 08:50:57.695603 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:50:57 crc kubenswrapper[4967]: E0120 08:50:57.695810 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.701676 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 21:44:02.123160076 +0000 UTC Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.705606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.705696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.705715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.705748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.705770 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.808179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.808235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.808252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.808276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.808296 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.911424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.911468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.911477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.911493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:57 crc kubenswrapper[4967]: I0120 08:50:57.911502 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:57Z","lastTransitionTime":"2026-01-20T08:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.014383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.014791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.014944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.015076 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.015191 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.118068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.118102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.118112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.118127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.118138 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.221351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.221411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.221429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.221455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.221473 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.324138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.324210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.324225 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.324242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.324254 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.427413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.427454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.427464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.427480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.427494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.530261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.530292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.530299 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.530311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.530321 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.633527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.633644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.633674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.633705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.633757 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.702350 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 22:36:11.259061226 +0000 UTC Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.736512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.736547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.736556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.736569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.736577 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.839461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.839538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.839562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.839590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.839654 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.942720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.942769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.942778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.942793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:58 crc kubenswrapper[4967]: I0120 08:50:58.942802 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:58Z","lastTransitionTime":"2026-01-20T08:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.044926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.045010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.045035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.045065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.045084 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.147828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.147903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.147929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.147960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.147977 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.250131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.250186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.250206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.250229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.250243 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.353222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.353296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.353319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.353348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.353368 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.456149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.456224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.456247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.456277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.456298 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.559030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.559105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.559139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.559169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.559192 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.662439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.662499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.662515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.662540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.662556 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.694222 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.694310 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:50:59 crc kubenswrapper[4967]: E0120 08:50:59.694433 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.694474 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.694485 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:50:59 crc kubenswrapper[4967]: E0120 08:50:59.695056 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:50:59 crc kubenswrapper[4967]: E0120 08:50:59.695285 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:50:59 crc kubenswrapper[4967]: E0120 08:50:59.695197 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.702833 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 01:25:35.012555877 +0000 UTC Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.765549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.765600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.765660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.765699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.765717 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.868729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.868793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.868811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.868834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.868852 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.971361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.971437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.971449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.971467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:50:59 crc kubenswrapper[4967]: I0120 08:50:59.971480 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:50:59Z","lastTransitionTime":"2026-01-20T08:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.074310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.074367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.074385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.074410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.074427 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.176983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.177042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.177059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.177080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.177095 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.279460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.279496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.279505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.279520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.279529 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.382690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.382771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.382794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.382822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.382841 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.485560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.485933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.486037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.486145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.486261 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.589850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.589893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.589903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.589919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.589932 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.692956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.693017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.693046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.693075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.693099 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.703521 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 06:52:41.398931799 +0000 UTC Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.795774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.795848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.795871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.795902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.795924 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.899126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.899192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.899215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.899250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:00 crc kubenswrapper[4967]: I0120 08:51:00.899273 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:00Z","lastTransitionTime":"2026-01-20T08:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.002507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.002575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.002596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.002655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.002677 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.104926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.105144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.105289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.105430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.105541 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.208602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.208696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.208716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.208741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.208791 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.311080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.311159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.311183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.311214 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.311242 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.413257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.413296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.413306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.413320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.413329 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.515435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.515486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.515504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.515535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.515551 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.618359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.618816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.619460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.620052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.620454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.693366 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:01 crc kubenswrapper[4967]: E0120 08:51:01.693533 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.693870 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.693901 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:01 crc kubenswrapper[4967]: E0120 08:51:01.693977 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.694158 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:01 crc kubenswrapper[4967]: E0120 08:51:01.694251 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:01 crc kubenswrapper[4967]: E0120 08:51:01.694511 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.704451 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 10:33:32.697477481 +0000 UTC Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.723740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.723971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.724124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.724311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.724499 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.828008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.828079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.828100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.828122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.828140 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.930654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.930699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.930713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.930734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:01 crc kubenswrapper[4967]: I0120 08:51:01.930747 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:01Z","lastTransitionTime":"2026-01-20T08:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.033563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.033684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.033709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.033738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.033763 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.137183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.137245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.137260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.137281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.137296 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.240421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.240910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.241095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.241286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.241501 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.324831 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:02 crc kubenswrapper[4967]: E0120 08:51:02.324990 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:51:02 crc kubenswrapper[4967]: E0120 08:51:02.325080 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs podName:ebe5f8d7-df5d-46aa-b33e-c3d6768bb245 nodeName:}" failed. No retries permitted until 2026-01-20 08:52:06.325052177 +0000 UTC m=+160.910352384 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs") pod "network-metrics-daemon-lxp6p" (UID: "ebe5f8d7-df5d-46aa-b33e-c3d6768bb245") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.344548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.344590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.344599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.344630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.344639 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.446913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.446963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.446970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.446983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.446992 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.549364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.549444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.549462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.549488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.549505 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.651894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.651955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.651972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.651995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.652013 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.705473 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:25:41.369121897 +0000 UTC Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.755472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.755555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.755577 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.755601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.755656 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.858891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.858935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.858954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.858977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.858995 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.961828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.961875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.961886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.961901 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:02 crc kubenswrapper[4967]: I0120 08:51:02.961910 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:02Z","lastTransitionTime":"2026-01-20T08:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.063701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.063740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.063748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.063763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.063772 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.165746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.165794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.165805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.165821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.165832 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.268678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.268709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.268720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.268743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.268754 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.370892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.370939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.370948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.370963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.370974 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.472545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.472591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.472603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.472641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.472656 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.575940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.576011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.576033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.576061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.576081 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.678533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.678598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.678665 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.678685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.678696 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.694106 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.694145 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.694338 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.694340 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:03 crc kubenswrapper[4967]: E0120 08:51:03.694424 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:03 crc kubenswrapper[4967]: E0120 08:51:03.694569 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:03 crc kubenswrapper[4967]: E0120 08:51:03.694721 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:03 crc kubenswrapper[4967]: E0120 08:51:03.694837 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.706458 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:03:43.164171742 +0000 UTC Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.781354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.781387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.781395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.781410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.781419 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.884155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.884222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.884239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.884263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.884280 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.987222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.987289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.987314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.987342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:03 crc kubenswrapper[4967]: I0120 08:51:03.987362 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:03Z","lastTransitionTime":"2026-01-20T08:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.090072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.090125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.090137 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.090159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.090171 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.197638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.197683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.197695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.197715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.197727 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.300136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.300196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.300208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.300231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.300245 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.402631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.402680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.402688 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.402702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.402713 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.505191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.505226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.505237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.505253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.505263 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.607236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.607275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.607289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.607303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.607313 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.706743 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 14:23:04.010310521 +0000 UTC Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.709560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.709590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.709599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.709631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.709640 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.811835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.811866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.811881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.811893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.811902 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.913591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.913691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.913713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.913971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:04 crc kubenswrapper[4967]: I0120 08:51:04.913994 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:04Z","lastTransitionTime":"2026-01-20T08:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.016856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.016923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.016936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.016954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.016966 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.119560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.119830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.119859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.119887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.119907 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.223161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.223232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.223251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.223276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.223295 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.325967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.326012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.326030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.326048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.326061 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.428121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.428157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.428168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.428185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.428197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.531000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.531041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.531052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.531068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.531079 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.632902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.632967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.632991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.633019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.633040 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.693776 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.693830 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.693837 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:05 crc kubenswrapper[4967]: E0120 08:51:05.693937 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:05 crc kubenswrapper[4967]: E0120 08:51:05.694063 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:05 crc kubenswrapper[4967]: E0120 08:51:05.694165 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.693807 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:05 crc kubenswrapper[4967]: E0120 08:51:05.694541 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.707221 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 10:06:56.697544722 +0000 UTC Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.712576 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=37.712565247 podStartE2EDuration="37.712565247s" podCreationTimestamp="2026-01-20 08:50:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.712518356 +0000 UTC m=+100.297818573" watchObservedRunningTime="2026-01-20 08:51:05.712565247 +0000 UTC m=+100.297865454" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.735769 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-h8vz7" podStartSLOduration=81.735743134 podStartE2EDuration="1m21.735743134s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.735661422 +0000 UTC m=+100.320961719" watchObservedRunningTime="2026-01-20 08:51:05.735743134 +0000 UTC m=+100.321043381" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.736251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.736331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.736355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.736803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.736823 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.773835 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.773817532 podStartE2EDuration="1m22.773817532s" podCreationTimestamp="2026-01-20 08:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.753988527 +0000 UTC m=+100.339288754" watchObservedRunningTime="2026-01-20 08:51:05.773817532 +0000 UTC m=+100.359117749" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.840338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.840372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.840383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.840400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.840412 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.845040 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.84501795 podStartE2EDuration="52.84501795s" podCreationTimestamp="2026-01-20 08:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.843782787 +0000 UTC m=+100.429083014" watchObservedRunningTime="2026-01-20 08:51:05.84501795 +0000 UTC m=+100.430318197" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.845404 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-hlcgf" podStartSLOduration=81.845394841 podStartE2EDuration="1m21.845394841s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.8271828 +0000 UTC m=+100.412483017" watchObservedRunningTime="2026-01-20 08:51:05.845394841 +0000 UTC m=+100.430695078" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.905185 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-nxlff" podStartSLOduration=82.905166296 podStartE2EDuration="1m22.905166296s" podCreationTimestamp="2026-01-20 08:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.905071453 +0000 UTC m=+100.490371670" watchObservedRunningTime="2026-01-20 08:51:05.905166296 +0000 UTC m=+100.490466503" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.917278 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podStartSLOduration=81.917261638 podStartE2EDuration="1m21.917261638s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.91625076 +0000 UTC m=+100.501550977" watchObservedRunningTime="2026-01-20 08:51:05.917261638 +0000 UTC m=+100.502561845" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.928216 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mfb5x" podStartSLOduration=80.928197659 podStartE2EDuration="1m20.928197659s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.928024594 +0000 UTC m=+100.513324801" watchObservedRunningTime="2026-01-20 08:51:05.928197659 +0000 UTC m=+100.513497856" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.942786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.942834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.942845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.942862 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.942874 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:05Z","lastTransitionTime":"2026-01-20T08:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.954526 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.954506292 podStartE2EDuration="22.954506292s" podCreationTimestamp="2026-01-20 08:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.953861025 +0000 UTC m=+100.539161242" watchObservedRunningTime="2026-01-20 08:51:05.954506292 +0000 UTC m=+100.539806499" Jan 20 08:51:05 crc kubenswrapper[4967]: I0120 08:51:05.967776 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.967756917 podStartE2EDuration="1m19.967756917s" podCreationTimestamp="2026-01-20 08:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:05.967625783 +0000 UTC m=+100.552926000" watchObservedRunningTime="2026-01-20 08:51:05.967756917 +0000 UTC m=+100.553057134" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.037955 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6wm4t" podStartSLOduration=82.037938068 podStartE2EDuration="1m22.037938068s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:06.037265029 +0000 UTC m=+100.622565236" watchObservedRunningTime="2026-01-20 08:51:06.037938068 +0000 UTC m=+100.623238275" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.045063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.045108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.045122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.045139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.045152 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:06Z","lastTransitionTime":"2026-01-20T08:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.146871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.146898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.146906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.146918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.146926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:06Z","lastTransitionTime":"2026-01-20T08:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.179335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.179363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.179372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.179386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.179394 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-20T08:51:06Z","lastTransitionTime":"2026-01-20T08:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.224441 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7"] Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.224799 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.228739 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.228978 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.229017 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.230130 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.369063 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.369157 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5c418a6-9d6a-45ab-85c3-beefd8904b29-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.369407 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5c418a6-9d6a-45ab-85c3-beefd8904b29-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.369435 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.369459 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c418a6-9d6a-45ab-85c3-beefd8904b29-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470736 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5c418a6-9d6a-45ab-85c3-beefd8904b29-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470792 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470816 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c418a6-9d6a-45ab-85c3-beefd8904b29-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470854 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470886 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5c418a6-9d6a-45ab-85c3-beefd8904b29-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470981 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.470924 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b5c418a6-9d6a-45ab-85c3-beefd8904b29-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.471802 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b5c418a6-9d6a-45ab-85c3-beefd8904b29-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.480034 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c418a6-9d6a-45ab-85c3-beefd8904b29-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.489061 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5c418a6-9d6a-45ab-85c3-beefd8904b29-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6jh7\" (UID: \"b5c418a6-9d6a-45ab-85c3-beefd8904b29\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.537219 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.707808 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:13:32.927936677 +0000 UTC Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.707879 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 20 08:51:06 crc kubenswrapper[4967]: I0120 08:51:06.716433 4967 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.240581 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" event={"ID":"b5c418a6-9d6a-45ab-85c3-beefd8904b29","Type":"ContainerStarted","Data":"e6aef15587cce2216a20f72f27030ce0c2a51a7e30b485b95504ef6de2a1df73"} Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.240651 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" event={"ID":"b5c418a6-9d6a-45ab-85c3-beefd8904b29","Type":"ContainerStarted","Data":"a58503ac3681a8f2972f49e3030b341b42d9a12ae62e468beaf4ca0feaeaeaa5"} Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.693817 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.693907 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.693846 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:07 crc kubenswrapper[4967]: E0120 08:51:07.694036 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:07 crc kubenswrapper[4967]: E0120 08:51:07.694148 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:07 crc kubenswrapper[4967]: I0120 08:51:07.694208 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:07 crc kubenswrapper[4967]: E0120 08:51:07.694275 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:07 crc kubenswrapper[4967]: E0120 08:51:07.694328 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:08 crc kubenswrapper[4967]: I0120 08:51:08.694364 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:51:08 crc kubenswrapper[4967]: E0120 08:51:08.694543 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:51:09 crc kubenswrapper[4967]: I0120 08:51:09.693476 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:09 crc kubenswrapper[4967]: I0120 08:51:09.693552 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:09 crc kubenswrapper[4967]: I0120 08:51:09.693579 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:09 crc kubenswrapper[4967]: I0120 08:51:09.693635 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:09 crc kubenswrapper[4967]: E0120 08:51:09.693699 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:09 crc kubenswrapper[4967]: E0120 08:51:09.693789 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:09 crc kubenswrapper[4967]: E0120 08:51:09.693879 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:09 crc kubenswrapper[4967]: E0120 08:51:09.693986 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:11 crc kubenswrapper[4967]: I0120 08:51:11.693858 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:11 crc kubenswrapper[4967]: I0120 08:51:11.693904 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:11 crc kubenswrapper[4967]: I0120 08:51:11.693918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:11 crc kubenswrapper[4967]: E0120 08:51:11.694026 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:11 crc kubenswrapper[4967]: I0120 08:51:11.694087 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:11 crc kubenswrapper[4967]: E0120 08:51:11.694172 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:11 crc kubenswrapper[4967]: E0120 08:51:11.694276 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:11 crc kubenswrapper[4967]: E0120 08:51:11.694334 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:13 crc kubenswrapper[4967]: I0120 08:51:13.693819 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:13 crc kubenswrapper[4967]: I0120 08:51:13.693872 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:13 crc kubenswrapper[4967]: I0120 08:51:13.693892 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:13 crc kubenswrapper[4967]: E0120 08:51:13.693962 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:13 crc kubenswrapper[4967]: I0120 08:51:13.693822 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:13 crc kubenswrapper[4967]: E0120 08:51:13.694036 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:13 crc kubenswrapper[4967]: E0120 08:51:13.694132 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:13 crc kubenswrapper[4967]: E0120 08:51:13.694250 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:15 crc kubenswrapper[4967]: I0120 08:51:15.693402 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:15 crc kubenswrapper[4967]: I0120 08:51:15.693443 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:15 crc kubenswrapper[4967]: I0120 08:51:15.693403 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:15 crc kubenswrapper[4967]: I0120 08:51:15.694669 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:15 crc kubenswrapper[4967]: E0120 08:51:15.694656 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:15 crc kubenswrapper[4967]: E0120 08:51:15.694783 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:15 crc kubenswrapper[4967]: E0120 08:51:15.695129 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:15 crc kubenswrapper[4967]: E0120 08:51:15.695236 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:17 crc kubenswrapper[4967]: I0120 08:51:17.693565 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:17 crc kubenswrapper[4967]: I0120 08:51:17.693703 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:17 crc kubenswrapper[4967]: E0120 08:51:17.693779 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:17 crc kubenswrapper[4967]: I0120 08:51:17.693851 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:17 crc kubenswrapper[4967]: E0120 08:51:17.693978 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:17 crc kubenswrapper[4967]: I0120 08:51:17.694026 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:17 crc kubenswrapper[4967]: E0120 08:51:17.694159 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:17 crc kubenswrapper[4967]: E0120 08:51:17.694280 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.282007 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/1.log" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.282637 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/0.log" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.282688 4967 generic.go:334] "Generic (PLEG): container finished" podID="7ba8630f-92bc-4708-a722-a7e27c747073" containerID="124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41" exitCode=1 Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.282721 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerDied","Data":"124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41"} Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.282760 4967 scope.go:117] "RemoveContainer" containerID="650fb3ff2ec31986ebdbf641f24e436523951f0a86c03e6411422937e3ae074d" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.283381 4967 scope.go:117] "RemoveContainer" containerID="124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41" Jan 20 08:51:19 crc kubenswrapper[4967]: E0120 08:51:19.283712 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6wm4t_openshift-multus(7ba8630f-92bc-4708-a722-a7e27c747073)\"" pod="openshift-multus/multus-6wm4t" podUID="7ba8630f-92bc-4708-a722-a7e27c747073" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.305678 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6jh7" podStartSLOduration=95.305648785 podStartE2EDuration="1m35.305648785s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:07.256199432 +0000 UTC m=+101.841499659" watchObservedRunningTime="2026-01-20 08:51:19.305648785 +0000 UTC m=+113.890949042" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.693752 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.693799 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.693845 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:19 crc kubenswrapper[4967]: E0120 08:51:19.693948 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:19 crc kubenswrapper[4967]: E0120 08:51:19.694079 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:19 crc kubenswrapper[4967]: E0120 08:51:19.694191 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:19 crc kubenswrapper[4967]: I0120 08:51:19.694237 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:19 crc kubenswrapper[4967]: E0120 08:51:19.694327 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:20 crc kubenswrapper[4967]: I0120 08:51:20.286863 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/1.log" Jan 20 08:51:21 crc kubenswrapper[4967]: I0120 08:51:21.694170 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:21 crc kubenswrapper[4967]: I0120 08:51:21.694217 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:21 crc kubenswrapper[4967]: I0120 08:51:21.694229 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:21 crc kubenswrapper[4967]: E0120 08:51:21.694366 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:21 crc kubenswrapper[4967]: I0120 08:51:21.694388 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:21 crc kubenswrapper[4967]: E0120 08:51:21.694483 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:21 crc kubenswrapper[4967]: E0120 08:51:21.694586 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:21 crc kubenswrapper[4967]: E0120 08:51:21.694669 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:22 crc kubenswrapper[4967]: I0120 08:51:22.694285 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:51:22 crc kubenswrapper[4967]: E0120 08:51:22.695094 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fc9bg_openshift-ovn-kubernetes(9e9f22bb-157c-4556-9066-cd3cdacdf3af)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" Jan 20 08:51:23 crc kubenswrapper[4967]: I0120 08:51:23.694287 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:23 crc kubenswrapper[4967]: I0120 08:51:23.694408 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:23 crc kubenswrapper[4967]: I0120 08:51:23.694414 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:23 crc kubenswrapper[4967]: E0120 08:51:23.694505 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:23 crc kubenswrapper[4967]: I0120 08:51:23.694558 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:23 crc kubenswrapper[4967]: E0120 08:51:23.694763 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:23 crc kubenswrapper[4967]: E0120 08:51:23.695243 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:23 crc kubenswrapper[4967]: E0120 08:51:23.695303 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:25 crc kubenswrapper[4967]: I0120 08:51:25.693815 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.694944 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:25 crc kubenswrapper[4967]: I0120 08:51:25.694976 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:25 crc kubenswrapper[4967]: I0120 08:51:25.695012 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.695109 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:25 crc kubenswrapper[4967]: I0120 08:51:25.695236 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.695309 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.696341 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.726748 4967 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 20 08:51:25 crc kubenswrapper[4967]: E0120 08:51:25.782322 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 20 08:51:27 crc kubenswrapper[4967]: I0120 08:51:27.693783 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:27 crc kubenswrapper[4967]: I0120 08:51:27.693801 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:27 crc kubenswrapper[4967]: I0120 08:51:27.693858 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:27 crc kubenswrapper[4967]: E0120 08:51:27.693996 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:27 crc kubenswrapper[4967]: I0120 08:51:27.694060 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:27 crc kubenswrapper[4967]: E0120 08:51:27.694077 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:27 crc kubenswrapper[4967]: E0120 08:51:27.694278 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:27 crc kubenswrapper[4967]: E0120 08:51:27.694386 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:29 crc kubenswrapper[4967]: I0120 08:51:29.694201 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:29 crc kubenswrapper[4967]: I0120 08:51:29.694271 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:29 crc kubenswrapper[4967]: E0120 08:51:29.694406 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:29 crc kubenswrapper[4967]: I0120 08:51:29.694469 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:29 crc kubenswrapper[4967]: I0120 08:51:29.694590 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:29 crc kubenswrapper[4967]: E0120 08:51:29.694670 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:29 crc kubenswrapper[4967]: E0120 08:51:29.694780 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:29 crc kubenswrapper[4967]: E0120 08:51:29.694920 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:30 crc kubenswrapper[4967]: I0120 08:51:30.694087 4967 scope.go:117] "RemoveContainer" containerID="124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41" Jan 20 08:51:30 crc kubenswrapper[4967]: E0120 08:51:30.783441 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.326695 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/1.log" Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.326796 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerStarted","Data":"4238c896cd7352173c8ecf22ea04a431be302611922f554cfa608261d0b94127"} Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.694049 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.694356 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.694072 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:31 crc kubenswrapper[4967]: E0120 08:51:31.694430 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:31 crc kubenswrapper[4967]: I0120 08:51:31.694516 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:31 crc kubenswrapper[4967]: E0120 08:51:31.694713 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:31 crc kubenswrapper[4967]: E0120 08:51:31.694912 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:31 crc kubenswrapper[4967]: E0120 08:51:31.694986 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:33 crc kubenswrapper[4967]: I0120 08:51:33.693887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:33 crc kubenswrapper[4967]: I0120 08:51:33.694282 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:33 crc kubenswrapper[4967]: E0120 08:51:33.694324 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:33 crc kubenswrapper[4967]: I0120 08:51:33.694385 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:33 crc kubenswrapper[4967]: I0120 08:51:33.694425 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:33 crc kubenswrapper[4967]: E0120 08:51:33.694559 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:33 crc kubenswrapper[4967]: E0120 08:51:33.694776 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:33 crc kubenswrapper[4967]: E0120 08:51:33.694829 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:34 crc kubenswrapper[4967]: I0120 08:51:34.695413 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.345996 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/3.log" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.348814 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerStarted","Data":"0ced0c42ffa0f0de4e4877f02fe69cf0b87ba9b8bd3cdfdd007227c3f57bad64"} Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.349126 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.374866 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podStartSLOduration=111.374847799 podStartE2EDuration="1m51.374847799s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:35.37451885 +0000 UTC m=+129.959819057" watchObservedRunningTime="2026-01-20 08:51:35.374847799 +0000 UTC m=+129.960148006" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.527700 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lxp6p"] Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.527843 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:35 crc kubenswrapper[4967]: E0120 08:51:35.527932 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.694673 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:35 crc kubenswrapper[4967]: E0120 08:51:35.694777 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.694977 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:35 crc kubenswrapper[4967]: E0120 08:51:35.695032 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:35 crc kubenswrapper[4967]: I0120 08:51:35.695695 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:35 crc kubenswrapper[4967]: E0120 08:51:35.695819 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:35 crc kubenswrapper[4967]: E0120 08:51:35.784762 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 20 08:51:36 crc kubenswrapper[4967]: I0120 08:51:36.694379 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:36 crc kubenswrapper[4967]: E0120 08:51:36.695010 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:37 crc kubenswrapper[4967]: I0120 08:51:37.694216 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:37 crc kubenswrapper[4967]: I0120 08:51:37.694217 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:37 crc kubenswrapper[4967]: E0120 08:51:37.694434 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:37 crc kubenswrapper[4967]: E0120 08:51:37.694527 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:37 crc kubenswrapper[4967]: I0120 08:51:37.694342 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:37 crc kubenswrapper[4967]: E0120 08:51:37.694731 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:38 crc kubenswrapper[4967]: I0120 08:51:38.693282 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:38 crc kubenswrapper[4967]: E0120 08:51:38.693519 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:39 crc kubenswrapper[4967]: I0120 08:51:39.693832 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:39 crc kubenswrapper[4967]: I0120 08:51:39.693955 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:39 crc kubenswrapper[4967]: E0120 08:51:39.694139 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 20 08:51:39 crc kubenswrapper[4967]: E0120 08:51:39.694435 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 20 08:51:39 crc kubenswrapper[4967]: I0120 08:51:39.694795 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:39 crc kubenswrapper[4967]: E0120 08:51:39.694886 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 20 08:51:40 crc kubenswrapper[4967]: I0120 08:51:40.693814 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:40 crc kubenswrapper[4967]: E0120 08:51:40.693973 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lxp6p" podUID="ebe5f8d7-df5d-46aa-b33e-c3d6768bb245" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.693608 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.693945 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.696094 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.696416 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.696659 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.696707 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 20 08:51:41 crc kubenswrapper[4967]: I0120 08:51:41.698900 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:42 crc kubenswrapper[4967]: I0120 08:51:42.693405 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:51:42 crc kubenswrapper[4967]: I0120 08:51:42.696642 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 20 08:51:42 crc kubenswrapper[4967]: I0120 08:51:42.697338 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.116708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.172395 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-plddg"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.174088 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.183442 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gpj5p"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.184444 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.198708 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.199114 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.199304 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.203195 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.203419 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.203949 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.204284 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.204603 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.205001 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.204956 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.218035 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.221507 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.221968 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222238 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222376 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222465 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222307 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222713 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.222275 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.223327 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.223607 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.230696 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.230903 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231106 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231171 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231253 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231546 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231569 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231574 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231656 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231743 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231792 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231865 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.231867 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.235038 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h6s8s"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.235626 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.236092 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.238277 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.238436 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qfnnr"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.238540 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.238926 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.239575 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.238962 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.239033 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.239455 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.249890 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.250671 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-57z2c"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.251261 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.251267 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.251887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.252810 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7pllc"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.253260 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.257511 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x9x4t"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.258168 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.258960 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.259294 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.259425 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.259664 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.259714 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.263841 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.264129 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.264312 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.264337 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.264854 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.265700 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.265740 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.266478 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.267057 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.267156 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-z2ps9"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.267261 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.267569 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.267754 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.268130 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.268267 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.268440 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.268203 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.268636 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.282885 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.284361 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.289249 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.289573 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.289596 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc"] Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.290709 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.291005 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.292374 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.292538 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.292646 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.293991 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.294261 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 20 08:51:47 crc kubenswrapper[4967]: I0120 08:51:47.296099 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.100772 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.101251 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.101387 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.101989 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.102237 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.102446 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.102710 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.103138 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.103577 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.103648 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.103874 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104190 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104265 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104706 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104780 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104806 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104796 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104877 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104928 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.104963 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.105020 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.105316 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.105817 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.105527 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.105556 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.106692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.106798 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107337 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107441 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107477 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-config\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107510 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-image-import-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107763 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107813 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdnrr\" (UniqueName: \"kubernetes.io/projected/76b01181-c4c3-4ae6-8768-965aa2a235a3-kube-api-access-bdnrr\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107855 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtn6d\" (UniqueName: \"kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107888 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107917 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107919 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.107983 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-trusted-ca\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108043 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6f75790-6be7-4f4c-ae17-5258a715666e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108074 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf2pj\" (UniqueName: \"kubernetes.io/projected/340acfd5-f5f9-4db9-8d90-d72a1ccac052-kube-api-access-pf2pj\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108100 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108125 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-config\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108150 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108173 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108195 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108223 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108247 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-auth-proxy-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108276 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-client\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108299 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz4zx\" (UniqueName: \"kubernetes.io/projected/f6f75790-6be7-4f4c-ae17-5258a715666e-kube-api-access-sz4zx\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108330 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108376 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108400 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkpjt\" (UniqueName: \"kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108422 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-encryption-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108472 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108529 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-config\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108570 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108598 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e9f372-2ab2-4caa-a9e2-4c563a446126-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108637 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108817 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108858 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108876 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108895 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.108961 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/340acfd5-f5f9-4db9-8d90-d72a1ccac052-machine-approver-tls\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109011 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbzn9\" (UniqueName: \"kubernetes.io/projected/a3e9f372-2ab2-4caa-a9e2-4c563a446126-kube-api-access-jbzn9\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109051 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8db30b00-67b2-4713-bb9e-1b853cbf08e1-serving-cert\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109088 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-serving-cert\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109130 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109169 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109252 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109283 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109349 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjlsg\" (UniqueName: \"kubernetes.io/projected/8db30b00-67b2-4713-bb9e-1b853cbf08e1-kube-api-access-fjlsg\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109386 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit-dir\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109407 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkbgz\" (UniqueName: \"kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109431 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109451 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scl8g\" (UniqueName: \"kubernetes.io/projected/b9553608-55a3-4c39-944c-385e329c79b6-kube-api-access-scl8g\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109469 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-node-pullsecrets\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109470 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109532 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109642 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109677 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9553608-55a3-4c39-944c-385e329c79b6-serving-cert\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109698 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109703 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2wt5\" (UniqueName: \"kubernetes.io/projected/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-kube-api-access-j2wt5\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109854 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-images\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109884 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109916 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.109963 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.110356 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.110598 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.110724 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.111220 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.111643 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.116870 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.117836 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.121080 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.121722 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.121860 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122017 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122029 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122160 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122395 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122781 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.122970 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-75ccq"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123183 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123236 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123353 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123693 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123860 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123733 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.123751 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.124084 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.125303 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.125690 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.127322 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.130774 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.133048 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.136749 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-pm5cd"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.137194 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.137879 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.138206 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.141060 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.141332 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.146421 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.147079 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.147499 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-plddg"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.149010 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.149717 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.150839 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.151284 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152119 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152188 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152253 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152119 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152384 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152506 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152645 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.152775 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.153230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.153282 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.153553 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.153935 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.154101 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.154399 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.154431 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.154435 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.154587 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.155291 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.155470 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.155716 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.156668 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.156778 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.156941 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.157035 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.157258 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.157935 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.158557 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.158629 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.159421 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.159478 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.159589 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.161394 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.161494 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.161801 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.161841 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.162111 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.162315 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.162507 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.163039 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.163430 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.164083 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.165298 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.165715 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.166501 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.168820 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.170132 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9fk7s"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.170942 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.173049 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.173950 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.174924 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.175373 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.178067 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.180411 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.181521 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.181637 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4fzjb"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.182408 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.183526 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h6s8s"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.186105 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gpj5p"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.186166 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z2ps9"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.187458 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.189067 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xp9k6"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.195780 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qfnnr"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.195883 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.196964 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.201464 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.204200 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7pllc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210281 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-audit-policies\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210509 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae790671-4e34-4dba-86ed-6d2851366835-audit-dir\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210643 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-oauth-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210762 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-images\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210884 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.210971 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9553608-55a3-4c39-944c-385e329c79b6-serving-cert\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211045 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211117 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211202 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211306 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-service-ca\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211431 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkrnv\" (UniqueName: \"kubernetes.io/projected/c6512677-6466-4738-a2d8-726f0f85a9fd-kube-api-access-mkrnv\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211528 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2wt5\" (UniqueName: \"kubernetes.io/projected/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-kube-api-access-j2wt5\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211859 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-images\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.213360 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.214002 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e05474a9-df95-4593-a3f5-ae1e8cb3785f-metrics-tls\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.214099 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-encryption-config\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.212695 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.213663 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.213763 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.213919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-images\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.211802 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.214908 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.214780 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.213308 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.214954 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215124 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3d6730d-549e-4a20-af1a-3096aab8c184-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215179 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-config\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215214 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-image-import-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215248 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjpl5\" (UniqueName: \"kubernetes.io/projected/ae790671-4e34-4dba-86ed-6d2851366835-kube-api-access-fjpl5\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215276 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6512677-6466-4738-a2d8-726f0f85a9fd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215303 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215330 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215379 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdnrr\" (UniqueName: \"kubernetes.io/projected/76b01181-c4c3-4ae6-8768-965aa2a235a3-kube-api-access-bdnrr\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215410 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215434 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215457 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtn6d\" (UniqueName: \"kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215514 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215537 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215567 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-trusted-ca\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215600 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xnkz\" (UniqueName: \"kubernetes.io/projected/8c931549-0309-4d10-8908-7c972199f0df-kube-api-access-7xnkz\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215656 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d6730d-549e-4a20-af1a-3096aab8c184-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215685 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-serving-cert\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215729 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6f75790-6be7-4f4c-ae17-5258a715666e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215783 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215813 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf2pj\" (UniqueName: \"kubernetes.io/projected/340acfd5-f5f9-4db9-8d90-d72a1ccac052-kube-api-access-pf2pj\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215843 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215876 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-config\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215907 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215926 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215954 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.215988 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216010 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216034 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-auth-proxy-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216060 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-client\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216086 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6512677-6466-4738-a2d8-726f0f85a9fd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216109 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz4zx\" (UniqueName: \"kubernetes.io/projected/f6f75790-6be7-4f4c-ae17-5258a715666e-kube-api-access-sz4zx\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216127 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216156 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216181 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86h5h\" (UniqueName: \"kubernetes.io/projected/e05474a9-df95-4593-a3f5-ae1e8cb3785f-kube-api-access-86h5h\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216199 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d6730d-549e-4a20-af1a-3096aab8c184-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216199 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216219 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q42g\" (UniqueName: \"kubernetes.io/projected/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-kube-api-access-6q42g\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216298 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-oauth-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216342 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216368 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkpjt\" (UniqueName: \"kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216391 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-encryption-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216417 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216441 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k26zg\" (UniqueName: \"kubernetes.io/projected/ace51544-8fe2-4352-bcd8-93611e156b80-kube-api-access-k26zg\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216481 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-config\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216505 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-config\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216515 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6f75790-6be7-4f4c-ae17-5258a715666e-config\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216564 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e9f372-2ab2-4caa-a9e2-4c563a446126-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216573 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-image-import-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216586 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216628 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216655 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216725 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/340acfd5-f5f9-4db9-8d90-d72a1ccac052-machine-approver-tls\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216752 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216777 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-etcd-client\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216800 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216827 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snks4\" (UniqueName: \"kubernetes.io/projected/63125d8f-2be4-492c-8b42-057c5035715a-kube-api-access-snks4\") pod \"downloads-7954f5f757-z2ps9\" (UID: \"63125d8f-2be4-492c-8b42-057c5035715a\") " pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216853 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbzn9\" (UniqueName: \"kubernetes.io/projected/a3e9f372-2ab2-4caa-a9e2-4c563a446126-kube-api-access-jbzn9\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216881 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8db30b00-67b2-4713-bb9e-1b853cbf08e1-serving-cert\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216937 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-serving-cert\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216959 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-trusted-ca-bundle\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.216987 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217015 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217031 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217056 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhmg5\" (UniqueName: \"kubernetes.io/projected/90b763ed-98ef-4961-86bc-a624bf3af585-kube-api-access-vhmg5\") pod \"migrator-59844c95c7-9ssxc\" (UID: \"90b763ed-98ef-4961-86bc-a624bf3af585\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217085 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217110 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217134 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217159 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c931549-0309-4d10-8908-7c972199f0df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217189 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217215 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit-dir\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217237 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvpst\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-kube-api-access-pvpst\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217257 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ace51544-8fe2-4352-bcd8-93611e156b80-proxy-tls\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217282 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjlsg\" (UniqueName: \"kubernetes.io/projected/8db30b00-67b2-4713-bb9e-1b853cbf08e1-kube-api-access-fjlsg\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217306 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-node-pullsecrets\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c931549-0309-4d10-8908-7c972199f0df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217356 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkbgz\" (UniqueName: \"kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217379 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scl8g\" (UniqueName: \"kubernetes.io/projected/b9553608-55a3-4c39-944c-385e329c79b6-kube-api-access-scl8g\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.217669 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.218147 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.218244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.219444 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.220405 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.221267 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.220656 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.220476 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.222302 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.222317 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.222422 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.222487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9553608-55a3-4c39-944c-385e329c79b6-serving-cert\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.222861 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9553608-55a3-4c39-944c-385e329c79b6-config\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.223206 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.223459 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-serving-ca\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.223961 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.224522 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/340acfd5-f5f9-4db9-8d90-d72a1ccac052-auth-proxy-config\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.226869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.227155 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-etcd-client\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.227165 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.227485 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-encryption-config\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.227559 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-node-pullsecrets\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231371 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/340acfd5-f5f9-4db9-8d90-d72a1ccac052-machine-approver-tls\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231428 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231889 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231909 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231748 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.232305 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e9f372-2ab2-4caa-a9e2-4c563a446126-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.232512 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.232703 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76b01181-c4c3-4ae6-8768-965aa2a235a3-audit-dir\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233066 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-config\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.231556 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8db30b00-67b2-4713-bb9e-1b853cbf08e1-serving-cert\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233259 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233298 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233402 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233786 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.233896 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.234006 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db30b00-67b2-4713-bb9e-1b853cbf08e1-trusted-ca\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.234937 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.235030 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.235210 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b01181-c4c3-4ae6-8768-965aa2a235a3-serving-cert\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.236088 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.236218 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.236305 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6f75790-6be7-4f4c-ae17-5258a715666e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.237076 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.237328 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txfwl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.237908 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.237363 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.238559 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s9jgl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.238902 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.239096 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.239721 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.239743 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.240604 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x9x4t"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.240942 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.242363 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.243629 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.243867 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.246757 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-57z2c"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.246834 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4fzjb"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.247031 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-75ccq"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.248328 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.249371 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.250723 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.252665 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.252706 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9fk7s"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.253901 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s9jgl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.255172 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xp9k6"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.256317 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.256581 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.258266 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txfwl"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.259234 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.260643 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vc6gw"] Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.263174 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.277764 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.296753 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.318898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k26zg\" (UniqueName: \"kubernetes.io/projected/ace51544-8fe2-4352-bcd8-93611e156b80-kube-api-access-k26zg\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.318947 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-config\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.318976 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.318993 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snks4\" (UniqueName: \"kubernetes.io/projected/63125d8f-2be4-492c-8b42-057c5035715a-kube-api-access-snks4\") pod \"downloads-7954f5f757-z2ps9\" (UID: \"63125d8f-2be4-492c-8b42-057c5035715a\") " pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319014 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319033 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-etcd-client\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319047 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319048 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319068 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-trusted-ca-bundle\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319087 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhmg5\" (UniqueName: \"kubernetes.io/projected/90b763ed-98ef-4961-86bc-a624bf3af585-kube-api-access-vhmg5\") pod \"migrator-59844c95c7-9ssxc\" (UID: \"90b763ed-98ef-4961-86bc-a624bf3af585\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319110 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c931549-0309-4d10-8908-7c972199f0df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319127 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319153 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvpst\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-kube-api-access-pvpst\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319170 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ace51544-8fe2-4352-bcd8-93611e156b80-proxy-tls\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319195 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c931549-0309-4d10-8908-7c972199f0df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319218 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-audit-policies\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319235 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae790671-4e34-4dba-86ed-6d2851366835-audit-dir\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319250 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-oauth-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319269 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-images\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319287 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-service-ca\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkrnv\" (UniqueName: \"kubernetes.io/projected/c6512677-6466-4738-a2d8-726f0f85a9fd-kube-api-access-mkrnv\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319328 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e05474a9-df95-4593-a3f5-ae1e8cb3785f-metrics-tls\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319343 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-encryption-config\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319360 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3d6730d-549e-4a20-af1a-3096aab8c184-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjpl5\" (UniqueName: \"kubernetes.io/projected/ae790671-4e34-4dba-86ed-6d2851366835-kube-api-access-fjpl5\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6512677-6466-4738-a2d8-726f0f85a9fd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319417 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319448 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319471 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xnkz\" (UniqueName: \"kubernetes.io/projected/8c931549-0309-4d10-8908-7c972199f0df-kube-api-access-7xnkz\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319486 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d6730d-549e-4a20-af1a-3096aab8c184-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319502 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-serving-cert\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319526 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319551 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319635 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6512677-6466-4738-a2d8-726f0f85a9fd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319667 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86h5h\" (UniqueName: \"kubernetes.io/projected/e05474a9-df95-4593-a3f5-ae1e8cb3785f-kube-api-access-86h5h\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319683 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d6730d-549e-4a20-af1a-3096aab8c184-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319699 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q42g\" (UniqueName: \"kubernetes.io/projected/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-kube-api-access-6q42g\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.319716 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-oauth-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.320472 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-audit-policies\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.320728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae790671-4e34-4dba-86ed-6d2851366835-audit-dir\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.321514 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.321657 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-config\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.321866 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-service-ca\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.321904 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-trusted-ca-bundle\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.322256 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-oauth-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.322768 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3d6730d-549e-4a20-af1a-3096aab8c184-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.322872 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.324272 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae790671-4e34-4dba-86ed-6d2851366835-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.324251 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.324441 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6512677-6466-4738-a2d8-726f0f85a9fd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.325531 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-images\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.325548 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ace51544-8fe2-4352-bcd8-93611e156b80-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.326298 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-encryption-config\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.326665 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-oauth-config\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.326741 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-etcd-client\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.326882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.327094 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6512677-6466-4738-a2d8-726f0f85a9fd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.327639 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae790671-4e34-4dba-86ed-6d2851366835-serving-cert\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.327690 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ace51544-8fe2-4352-bcd8-93611e156b80-proxy-tls\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.328058 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e05474a9-df95-4593-a3f5-ae1e8cb3785f-metrics-tls\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.329111 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.329144 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3d6730d-549e-4a20-af1a-3096aab8c184-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.329346 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-console-serving-cert\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.338155 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.357955 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.377357 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.397653 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.416502 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.438385 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.457307 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.474094 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.474146 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.476931 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.497296 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.516978 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.537398 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.564466 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.577916 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.616898 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.636467 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.657359 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.677554 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.698128 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.717435 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.724632 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c931549-0309-4d10-8908-7c972199f0df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.736808 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.742084 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c931549-0309-4d10-8908-7c972199f0df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.758173 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.778027 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.798174 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.817470 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.838517 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.857816 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.878536 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.897747 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.917889 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.938313 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.957962 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.978752 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 20 08:51:48 crc kubenswrapper[4967]: I0120 08:51:48.997970 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.017457 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.037583 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.118338 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtn6d\" (UniqueName: \"kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d\") pod \"route-controller-manager-6576b87f9c-n544b\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.118706 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2wt5\" (UniqueName: \"kubernetes.io/projected/8155bf3a-4152-4f9c-a9aa-7e3abcf84912-kube-api-access-j2wt5\") pod \"openshift-config-operator-7777fb866f-mgjxq\" (UID: \"8155bf3a-4152-4f9c-a9aa-7e3abcf84912\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.137638 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scl8g\" (UniqueName: \"kubernetes.io/projected/b9553608-55a3-4c39-944c-385e329c79b6-kube-api-access-scl8g\") pod \"authentication-operator-69f744f599-qfnnr\" (UID: \"b9553608-55a3-4c39-944c-385e329c79b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.154940 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdnrr\" (UniqueName: \"kubernetes.io/projected/76b01181-c4c3-4ae6-8768-965aa2a235a3-kube-api-access-bdnrr\") pod \"apiserver-76f77b778f-plddg\" (UID: \"76b01181-c4c3-4ae6-8768-965aa2a235a3\") " pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.177703 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkpjt\" (UniqueName: \"kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt\") pod \"oauth-openshift-558db77b4-57z2c\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.198639 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz4zx\" (UniqueName: \"kubernetes.io/projected/f6f75790-6be7-4f4c-ae17-5258a715666e-kube-api-access-sz4zx\") pod \"machine-api-operator-5694c8668f-gpj5p\" (UID: \"f6f75790-6be7-4f4c-ae17-5258a715666e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.215061 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjlsg\" (UniqueName: \"kubernetes.io/projected/8db30b00-67b2-4713-bb9e-1b853cbf08e1-kube-api-access-fjlsg\") pod \"console-operator-58897d9998-h6s8s\" (UID: \"8db30b00-67b2-4713-bb9e-1b853cbf08e1\") " pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.235927 4967 request.go:700] Waited for 1.00699756s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/serviceaccounts/openshift-controller-manager-sa/token Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.249422 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbzn9\" (UniqueName: \"kubernetes.io/projected/a3e9f372-2ab2-4caa-a9e2-4c563a446126-kube-api-access-jbzn9\") pod \"cluster-samples-operator-665b6dd947-tfmfv\" (UID: \"a3e9f372-2ab2-4caa-a9e2-4c563a446126\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.266521 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkbgz\" (UniqueName: \"kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz\") pod \"controller-manager-879f6c89f-fbdgc\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.277483 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.277830 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf2pj\" (UniqueName: \"kubernetes.io/projected/340acfd5-f5f9-4db9-8d90-d72a1ccac052-kube-api-access-pf2pj\") pod \"machine-approver-56656f9798-rm9cb\" (UID: \"340acfd5-f5f9-4db9-8d90-d72a1ccac052\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.297531 4967 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.313596 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.313760 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.318455 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.332200 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.338218 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.339923 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.342760 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.354340 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.357304 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.361377 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.365722 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.379021 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.379469 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.398399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.398818 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.417420 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.440520 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.462589 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.500866 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhmg5\" (UniqueName: \"kubernetes.io/projected/90b763ed-98ef-4961-86bc-a624bf3af585-kube-api-access-vhmg5\") pod \"migrator-59844c95c7-9ssxc\" (UID: \"90b763ed-98ef-4961-86bc-a624bf3af585\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.515491 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k26zg\" (UniqueName: \"kubernetes.io/projected/ace51544-8fe2-4352-bcd8-93611e156b80-kube-api-access-k26zg\") pod \"machine-config-operator-74547568cd-pm4ww\" (UID: \"ace51544-8fe2-4352-bcd8-93611e156b80\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.538005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.554472 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.555685 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qfnnr"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.564113 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvpst\" (UniqueName: \"kubernetes.io/projected/89ee25c2-17c2-4464-9dad-e9126c1fa0cf-kube-api-access-pvpst\") pod \"cluster-image-registry-operator-dc59b4c8b-q7jxn\" (UID: \"89ee25c2-17c2-4464-9dad-e9126c1fa0cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.574775 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3d6730d-549e-4a20-af1a-3096aab8c184-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cfrcf\" (UID: \"b3d6730d-549e-4a20-af1a-3096aab8c184\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.582278 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.595396 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fz8kc\" (UID: \"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.596638 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:51:49 crc kubenswrapper[4967]: W0120 08:51:49.605858 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9553608_55a3_4c39_944c_385e329c79b6.slice/crio-8b82770d8b9c91fe9b39305c2316e9604e817af07c885db6bc0aa25b306d4729 WatchSource:0}: Error finding container 8b82770d8b9c91fe9b39305c2316e9604e817af07c885db6bc0aa25b306d4729: Status 404 returned error can't find the container with id 8b82770d8b9c91fe9b39305c2316e9604e817af07c885db6bc0aa25b306d4729 Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.611370 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snks4\" (UniqueName: \"kubernetes.io/projected/63125d8f-2be4-492c-8b42-057c5035715a-kube-api-access-snks4\") pod \"downloads-7954f5f757-z2ps9\" (UID: \"63125d8f-2be4-492c-8b42-057c5035715a\") " pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.633463 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkrnv\" (UniqueName: \"kubernetes.io/projected/c6512677-6466-4738-a2d8-726f0f85a9fd-kube-api-access-mkrnv\") pod \"openshift-controller-manager-operator-756b6f6bc6-c9qgm\" (UID: \"c6512677-6466-4738-a2d8-726f0f85a9fd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.653379 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86h5h\" (UniqueName: \"kubernetes.io/projected/e05474a9-df95-4593-a3f5-ae1e8cb3785f-kube-api-access-86h5h\") pod \"dns-operator-744455d44c-x9x4t\" (UID: \"e05474a9-df95-4593-a3f5-ae1e8cb3785f\") " pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.674387 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q42g\" (UniqueName: \"kubernetes.io/projected/a00ea83f-d3ff-4f4a-a82f-83eaf4edd589-kube-api-access-6q42g\") pod \"console-f9d7485db-7pllc\" (UID: \"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589\") " pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.695251 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjpl5\" (UniqueName: \"kubernetes.io/projected/ae790671-4e34-4dba-86ed-6d2851366835-kube-api-access-fjpl5\") pod \"apiserver-7bbb656c7d-8lfg9\" (UID: \"ae790671-4e34-4dba-86ed-6d2851366835\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.709093 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.715539 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xnkz\" (UniqueName: \"kubernetes.io/projected/8c931549-0309-4d10-8908-7c972199f0df-kube-api-access-7xnkz\") pod \"kube-storage-version-migrator-operator-b67b599dd-rlj75\" (UID: \"8c931549-0309-4d10-8908-7c972199f0df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.723505 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.771899 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.779568 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.783759 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.787860 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.789207 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h6s8s"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.794948 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.802517 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.819953 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.839962 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.839998 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-client\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840027 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840065 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80899a1a-ffa8-4daa-9f19-d80ca422eea9-trusted-ca\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840089 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9fpr\" (UniqueName: \"kubernetes.io/projected/adcafa80-4a5f-4268-8a01-6284b6cf235c-kube-api-access-t9fpr\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840126 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-webhook-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840146 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4pfc\" (UniqueName: \"kubernetes.io/projected/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-kube-api-access-l4pfc\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840176 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a72e607-9834-431b-8095-193ae6452de3-config\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840238 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840261 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbv8z\" (UniqueName: \"kubernetes.io/projected/ced462f3-f1bf-440f-bd5a-6973fc418386-kube-api-access-wbv8z\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840301 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-serving-cert\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840324 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840344 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-config\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840364 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-stats-auth\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840401 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxbdx\" (UniqueName: \"kubernetes.io/projected/074019c5-93cc-496d-9ddd-49bac54ccfa5-kube-api-access-xxbdx\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840423 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns8ds\" (UniqueName: \"kubernetes.io/projected/55129ade-282b-431d-8e52-00e8396cfeaf-kube-api-access-ns8ds\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840452 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840545 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840598 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840668 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840693 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdn5q\" (UniqueName: \"kubernetes.io/projected/5e360baa-93ae-4aa9-b2db-b362d45b9fed-kube-api-access-hdn5q\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840719 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-cabundle\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840748 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a72e607-9834-431b-8095-193ae6452de3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840806 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-srv-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840826 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80899a1a-ffa8-4daa-9f19-d80ca422eea9-metrics-tls\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840847 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840878 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.840973 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e360baa-93ae-4aa9-b2db-b362d45b9fed-proxy-tls\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841001 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/074019c5-93cc-496d-9ddd-49bac54ccfa5-tmpfs\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841036 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72e607-9834-431b-8095-193ae6452de3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841055 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-key\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841078 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6449\" (UniqueName: \"kubernetes.io/projected/cc8e9166-fdeb-4308-a44e-e932397bfead-kube-api-access-x6449\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841118 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841142 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26vf2\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-kube-api-access-26vf2\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841164 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqsj9\" (UniqueName: \"kubernetes.io/projected/e4949fe9-aa28-4409-bf01-ed9247a422b3-kube-api-access-cqsj9\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841198 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqw5\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841220 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced462f3-f1bf-440f-bd5a-6973fc418386-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841254 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841276 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841295 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e360baa-93ae-4aa9-b2db-b362d45b9fed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841326 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvtd8\" (UniqueName: \"kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841346 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46hnt\" (UniqueName: \"kubernetes.io/projected/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-kube-api-access-46hnt\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841389 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs5pw\" (UniqueName: \"kubernetes.io/projected/cf0c805a-669e-422a-906a-917c335ed5ab-kube-api-access-vs5pw\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841408 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-default-certificate\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841441 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc8e9166-fdeb-4308-a44e-e932397bfead-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841503 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841526 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841559 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55129ade-282b-431d-8e52-00e8396cfeaf-serving-cert\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841582 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-service-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841602 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55129ade-282b-431d-8e52-00e8396cfeaf-config\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841671 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841691 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-srv-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841742 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841768 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x77n8\" (UniqueName: \"kubernetes.io/projected/27f66629-6bff-4ce5-820c-8aebe63600cc-kube-api-access-x77n8\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841808 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjzsv\" (UniqueName: \"kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841832 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8c5f152-a3c7-41b7-ae9d-2317eed21246-metrics-tls\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841853 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz85z\" (UniqueName: \"kubernetes.io/projected/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-kube-api-access-xz85z\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841878 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-metrics-certs\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk7gs\" (UniqueName: \"kubernetes.io/projected/b8c5f152-a3c7-41b7-ae9d-2317eed21246-kube-api-access-mk7gs\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841919 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841931 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gpj5p"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.841938 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/adcafa80-4a5f-4268-8a01-6284b6cf235c-service-ca-bundle\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.842055 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c5f152-a3c7-41b7-ae9d-2317eed21246-config-volume\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.873488 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-plddg"] Jan 20 08:51:49 crc kubenswrapper[4967]: E0120 08:51:49.877090 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.377072366 +0000 UTC m=+144.962372573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.881411 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.912029 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-57z2c"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.937561 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.942915 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.943131 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e360baa-93ae-4aa9-b2db-b362d45b9fed-proxy-tls\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.943163 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/074019c5-93cc-496d-9ddd-49bac54ccfa5-tmpfs\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.943185 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72e607-9834-431b-8095-193ae6452de3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.943201 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-key\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.943974 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6449\" (UniqueName: \"kubernetes.io/projected/cc8e9166-fdeb-4308-a44e-e932397bfead-kube-api-access-x6449\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944000 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944023 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a5f3fa-e077-4dae-be02-0a5967a2a15a-cert\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944061 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26vf2\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-kube-api-access-26vf2\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944081 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqsj9\" (UniqueName: \"kubernetes.io/projected/e4949fe9-aa28-4409-bf01-ed9247a422b3-kube-api-access-cqsj9\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944096 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqw5\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944111 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced462f3-f1bf-440f-bd5a-6973fc418386-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944128 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxnv\" (UniqueName: \"kubernetes.io/projected/03a5f3fa-e077-4dae-be02-0a5967a2a15a-kube-api-access-hcxnv\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944731 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.944788 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv"] Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945015 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8hks\" (UniqueName: \"kubernetes.io/projected/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-kube-api-access-w8hks\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945050 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945071 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e360baa-93ae-4aa9-b2db-b362d45b9fed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945116 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945135 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46hnt\" (UniqueName: \"kubernetes.io/projected/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-kube-api-access-46hnt\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945153 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945169 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvtd8\" (UniqueName: \"kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945188 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-default-certificate\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945213 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs5pw\" (UniqueName: \"kubernetes.io/projected/cf0c805a-669e-422a-906a-917c335ed5ab-kube-api-access-vs5pw\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945230 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc8e9166-fdeb-4308-a44e-e932397bfead-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945248 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945264 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945291 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55129ade-282b-431d-8e52-00e8396cfeaf-serving-cert\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945305 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55129ade-282b-431d-8e52-00e8396cfeaf-config\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945322 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-service-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945337 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-srv-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945356 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-socket-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945436 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x77n8\" (UniqueName: \"kubernetes.io/projected/27f66629-6bff-4ce5-820c-8aebe63600cc-kube-api-access-x77n8\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945462 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8c5f152-a3c7-41b7-ae9d-2317eed21246-metrics-tls\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945488 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz85z\" (UniqueName: \"kubernetes.io/projected/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-kube-api-access-xz85z\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945514 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjzsv\" (UniqueName: \"kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945548 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-csi-data-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945567 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk7gs\" (UniqueName: \"kubernetes.io/projected/b8c5f152-a3c7-41b7-ae9d-2317eed21246-kube-api-access-mk7gs\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945586 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-node-bootstrap-token\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945626 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945647 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/adcafa80-4a5f-4268-8a01-6284b6cf235c-service-ca-bundle\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945662 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-metrics-certs\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945706 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c5f152-a3c7-41b7-ae9d-2317eed21246-config-volume\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945737 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-client\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945768 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945792 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80899a1a-ffa8-4daa-9f19-d80ca422eea9-trusted-ca\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945814 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9fpr\" (UniqueName: \"kubernetes.io/projected/adcafa80-4a5f-4268-8a01-6284b6cf235c-kube-api-access-t9fpr\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-webhook-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945885 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4pfc\" (UniqueName: \"kubernetes.io/projected/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-kube-api-access-l4pfc\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945920 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a72e607-9834-431b-8095-193ae6452de3-config\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945944 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rpgn\" (UniqueName: \"kubernetes.io/projected/0173c2dd-2801-47b0-abb3-684fabcb1845-kube-api-access-5rpgn\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.945998 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbv8z\" (UniqueName: \"kubernetes.io/projected/ced462f3-f1bf-440f-bd5a-6973fc418386-kube-api-access-wbv8z\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946020 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-plugins-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946039 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-certs\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946063 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946079 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-serving-cert\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946112 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-config\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946129 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-stats-auth\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946147 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946166 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns8ds\" (UniqueName: \"kubernetes.io/projected/55129ade-282b-431d-8e52-00e8396cfeaf-kube-api-access-ns8ds\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946202 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxbdx\" (UniqueName: \"kubernetes.io/projected/074019c5-93cc-496d-9ddd-49bac54ccfa5-kube-api-access-xxbdx\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946260 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946279 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946300 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdn5q\" (UniqueName: \"kubernetes.io/projected/5e360baa-93ae-4aa9-b2db-b362d45b9fed-kube-api-access-hdn5q\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946337 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-mountpoint-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946354 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-cabundle\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946371 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a72e607-9834-431b-8095-193ae6452de3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946387 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-registration-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-srv-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946430 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80899a1a-ffa8-4daa-9f19-d80ca422eea9-metrics-tls\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.946481 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: E0120 08:51:49.946620 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.446580404 +0000 UTC m=+145.031880681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.949253 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/074019c5-93cc-496d-9ddd-49bac54ccfa5-tmpfs\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.951545 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.956092 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.957404 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.958186 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-service-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.958542 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.958785 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55129ade-282b-431d-8e52-00e8396cfeaf-config\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.959071 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-config\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.959074 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e360baa-93ae-4aa9-b2db-b362d45b9fed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.959985 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e360baa-93ae-4aa9-b2db-b362d45b9fed-proxy-tls\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.961734 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/adcafa80-4a5f-4268-8a01-6284b6cf235c-service-ca-bundle\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.961792 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.962192 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-ca\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.962707 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55129ade-282b-431d-8e52-00e8396cfeaf-serving-cert\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.963639 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-key\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.964069 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a72e607-9834-431b-8095-193ae6452de3-config\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.964089 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.965092 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.967096 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-default-certificate\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.967940 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.968985 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.969120 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80899a1a-ffa8-4daa-9f19-d80ca422eea9-trusted-ca\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.969206 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/27f66629-6bff-4ce5-820c-8aebe63600cc-signing-cabundle\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.969848 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.970421 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-etcd-client\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.970860 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.971073 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced462f3-f1bf-440f-bd5a-6973fc418386-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.971540 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8c5f152-a3c7-41b7-ae9d-2317eed21246-metrics-tls\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.973036 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-webhook-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.973443 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc8e9166-fdeb-4308-a44e-e932397bfead-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.973714 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.974171 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a72e607-9834-431b-8095-193ae6452de3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.975287 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/074019c5-93cc-496d-9ddd-49bac54ccfa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.977443 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-metrics-certs\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.977788 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.979729 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/adcafa80-4a5f-4268-8a01-6284b6cf235c-stats-auth\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.979803 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80899a1a-ffa8-4daa-9f19-d80ca422eea9-metrics-tls\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.982207 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf0c805a-669e-422a-906a-917c335ed5ab-srv-cert\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.983815 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4949fe9-aa28-4409-bf01-ed9247a422b3-serving-cert\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.983818 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.984061 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c5f152-a3c7-41b7-ae9d-2317eed21246-config-volume\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.984111 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:49 crc kubenswrapper[4967]: I0120 08:51:49.990276 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-srv-cert\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:49.997231 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqw5\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.020621 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqsj9\" (UniqueName: \"kubernetes.io/projected/e4949fe9-aa28-4409-bf01-ed9247a422b3-kube-api-access-cqsj9\") pod \"etcd-operator-b45778765-75ccq\" (UID: \"e4949fe9-aa28-4409-bf01-ed9247a422b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.022345 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6449\" (UniqueName: \"kubernetes.io/projected/cc8e9166-fdeb-4308-a44e-e932397bfead-kube-api-access-x6449\") pod \"control-plane-machine-set-operator-78cbb6b69f-5gqsg\" (UID: \"cc8e9166-fdeb-4308-a44e-e932397bfead\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.031507 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26vf2\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-kube-api-access-26vf2\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047550 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rpgn\" (UniqueName: \"kubernetes.io/projected/0173c2dd-2801-47b0-abb3-684fabcb1845-kube-api-access-5rpgn\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047586 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-plugins-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047600 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-certs\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047652 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047697 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-mountpoint-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047714 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-registration-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a5f3fa-e077-4dae-be02-0a5967a2a15a-cert\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047763 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxnv\" (UniqueName: \"kubernetes.io/projected/03a5f3fa-e077-4dae-be02-0a5967a2a15a-kube-api-access-hcxnv\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047779 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8hks\" (UniqueName: \"kubernetes.io/projected/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-kube-api-access-w8hks\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047847 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-socket-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047880 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-csi-data-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.047899 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-node-bootstrap-token\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.051909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-registration-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.052369 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-plugins-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.052426 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-mountpoint-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.052486 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.552455182 +0000 UTC m=+145.137755389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.052701 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz85z\" (UniqueName: \"kubernetes.io/projected/e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7-kube-api-access-xz85z\") pod \"openshift-apiserver-operator-796bbdcf4f-xrwv5\" (UID: \"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.052880 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-socket-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.052987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-csi-data-dir\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.062487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-node-bootstrap-token\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.063163 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a5f3fa-e077-4dae-be02-0a5967a2a15a-cert\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.076678 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7pllc"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.081552 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjzsv\" (UniqueName: \"kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv\") pod \"collect-profiles-29481645-m6h5v\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.082218 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0173c2dd-2801-47b0-abb3-684fabcb1845-certs\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.096119 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk7gs\" (UniqueName: \"kubernetes.io/projected/b8c5f152-a3c7-41b7-ae9d-2317eed21246-kube-api-access-mk7gs\") pod \"dns-default-xp9k6\" (UID: \"b8c5f152-a3c7-41b7-ae9d-2317eed21246\") " pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.144740 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" event={"ID":"b9553608-55a3-4c39-944c-385e329c79b6","Type":"ContainerStarted","Data":"b1f11a2fa4f4908eb20b2a8b93de5d8f76df24138537c0e1407e37220f9e8a9e"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.145071 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" event={"ID":"b9553608-55a3-4c39-944c-385e329c79b6","Type":"ContainerStarted","Data":"8b82770d8b9c91fe9b39305c2316e9604e817af07c885db6bc0aa25b306d4729"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.147213 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs5pw\" (UniqueName: \"kubernetes.io/projected/cf0c805a-669e-422a-906a-917c335ed5ab-kube-api-access-vs5pw\") pod \"olm-operator-6b444d44fb-f42q5\" (UID: \"cf0c805a-669e-422a-906a-917c335ed5ab\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.149406 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.149515 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.649496792 +0000 UTC m=+145.234796989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.149635 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" event={"ID":"f6f75790-6be7-4f4c-ae17-5258a715666e","Type":"ContainerStarted","Data":"af2e64171dafef33d8c29f5820d28997e41f11a71b9c41706ba839e537c62b24"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.149772 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.150114 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.65010095 +0000 UTC m=+145.235401157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.154384 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" event={"ID":"7db3ffb6-ef1c-432b-8235-5ef6d77525a4","Type":"ContainerStarted","Data":"98d92dbdbdbe21f3f06c41e2055aae90eb7ec4fb07eb5731e51c371ce2992529"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.161872 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" event={"ID":"8155bf3a-4152-4f9c-a9aa-7e3abcf84912","Type":"ContainerStarted","Data":"4e73df5b30bfba4569e66fb5073b42514cc3e9e55ab148587285b0d6249791f7"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.173668 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72e607-9834-431b-8095-193ae6452de3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hr6g\" (UID: \"0a72e607-9834-431b-8095-193ae6452de3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.174824 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.178071 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvtd8\" (UniqueName: \"kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8\") pod \"marketplace-operator-79b997595-ksh2s\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.184783 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.192564 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9fpr\" (UniqueName: \"kubernetes.io/projected/adcafa80-4a5f-4268-8a01-6284b6cf235c-kube-api-access-t9fpr\") pod \"router-default-5444994796-pm5cd\" (UID: \"adcafa80-4a5f-4268-8a01-6284b6cf235c\") " pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.194199 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.209402 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.214703 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" event={"ID":"293c612b-323d-4408-af28-b9f559766583","Type":"ContainerStarted","Data":"a5088aed30a9aa770894dffdfef93a31bd8134cbc129044bee6572b5537799ba"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.227994 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.237090 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.242739 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.250394 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.252027 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.752002652 +0000 UTC m=+145.337302919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.257839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" event={"ID":"340acfd5-f5f9-4db9-8d90-d72a1ccac052","Type":"ContainerStarted","Data":"6458a605f788fcf2a7ed50437249772209eecc053c937cd724c1eb32b5146418"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.257886 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" event={"ID":"340acfd5-f5f9-4db9-8d90-d72a1ccac052","Type":"ContainerStarted","Data":"76252ac2f830d127164215ed31a47526a47b11cf4cf7662f3fd1a7c67b9a2965"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.265087 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" event={"ID":"ac422dc4-1356-4cbe-b1ec-d99946de9a13","Type":"ContainerStarted","Data":"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.265133 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" event={"ID":"ac422dc4-1356-4cbe-b1ec-d99946de9a13","Type":"ContainerStarted","Data":"5765f60efb375c5a802963da33f7a96c172e76ce3745e0a28cf31287f953ed0c"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.273091 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.274050 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" event={"ID":"8db30b00-67b2-4713-bb9e-1b853cbf08e1","Type":"ContainerStarted","Data":"cdeecf769b2848d0e9697ff0f2dcdb4ea5b97d3527ef89fc97099b1296b58271"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.274706 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" event={"ID":"90b763ed-98ef-4961-86bc-a624bf3af585","Type":"ContainerStarted","Data":"e260bae99ae7a823f81fa2421c1021cb7b850e8253fe41ca70cf5721f9656a23"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.274928 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.275848 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" event={"ID":"ace51544-8fe2-4352-bcd8-93611e156b80","Type":"ContainerStarted","Data":"9ee9a2cc2381df00b3ce2ec398f6eae3619ce93a94b7bee1621f3d9d7b0699cc"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.276471 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-plddg" event={"ID":"76b01181-c4c3-4ae6-8768-965aa2a235a3","Type":"ContainerStarted","Data":"418ed6ed7ebb508e7b5a07551a952cf6abf0854c7b9bde4e5a60cbcc1ca3a876"} Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.276549 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.279241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxbdx\" (UniqueName: \"kubernetes.io/projected/074019c5-93cc-496d-9ddd-49bac54ccfa5-kube-api-access-xxbdx\") pod \"packageserver-d55dfcdfc-h56fr\" (UID: \"074019c5-93cc-496d-9ddd-49bac54ccfa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.283758 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns8ds\" (UniqueName: \"kubernetes.io/projected/55129ade-282b-431d-8e52-00e8396cfeaf-kube-api-access-ns8ds\") pod \"service-ca-operator-777779d784-d4bdk\" (UID: \"55129ade-282b-431d-8e52-00e8396cfeaf\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.293100 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.306141 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46hnt\" (UniqueName: \"kubernetes.io/projected/e38c077e-5d16-4200-a9bf-2eb9756fd5e7-kube-api-access-46hnt\") pod \"multus-admission-controller-857f4d67dd-4fzjb\" (UID: \"e38c077e-5d16-4200-a9bf-2eb9756fd5e7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.306250 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdn5q\" (UniqueName: \"kubernetes.io/projected/5e360baa-93ae-4aa9-b2db-b362d45b9fed-kube-api-access-hdn5q\") pod \"machine-config-controller-84d6567774-k8jmc\" (UID: \"5e360baa-93ae-4aa9-b2db-b362d45b9fed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.311934 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/80899a1a-ffa8-4daa-9f19-d80ca422eea9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xn64x\" (UID: \"80899a1a-ffa8-4daa-9f19-d80ca422eea9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.345244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbv8z\" (UniqueName: \"kubernetes.io/projected/ced462f3-f1bf-440f-bd5a-6973fc418386-kube-api-access-wbv8z\") pod \"package-server-manager-789f6589d5-pkrmf\" (UID: \"ced462f3-f1bf-440f-bd5a-6973fc418386\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.352207 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.352598 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.852585066 +0000 UTC m=+145.437885283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.365997 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4pfc\" (UniqueName: \"kubernetes.io/projected/cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8-kube-api-access-l4pfc\") pod \"catalog-operator-68c6474976-6p4jt\" (UID: \"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.370959 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x9x4t"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.375249 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x77n8\" (UniqueName: \"kubernetes.io/projected/27f66629-6bff-4ce5-820c-8aebe63600cc-kube-api-access-x77n8\") pod \"service-ca-9c57cc56f-9fk7s\" (UID: \"27f66629-6bff-4ce5-820c-8aebe63600cc\") " pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.404245 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rpgn\" (UniqueName: \"kubernetes.io/projected/0173c2dd-2801-47b0-abb3-684fabcb1845-kube-api-access-5rpgn\") pod \"machine-config-server-vc6gw\" (UID: \"0173c2dd-2801-47b0-abb3-684fabcb1845\") " pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.415751 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8hks\" (UniqueName: \"kubernetes.io/projected/a0d60d70-87ff-45b9-b7e3-053172ce2fe5-kube-api-access-w8hks\") pod \"csi-hostpathplugin-txfwl\" (UID: \"a0d60d70-87ff-45b9-b7e3-053172ce2fe5\") " pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.434151 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.434193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.449791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxnv\" (UniqueName: \"kubernetes.io/projected/03a5f3fa-e077-4dae-be02-0a5967a2a15a-kube-api-access-hcxnv\") pod \"ingress-canary-s9jgl\" (UID: \"03a5f3fa-e077-4dae-be02-0a5967a2a15a\") " pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.461431 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.461721 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.461743 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.961720669 +0000 UTC m=+145.547020876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.461901 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.462331 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:50.962316106 +0000 UTC m=+145.547616313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.518945 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.519859 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.527530 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.569044 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.570195 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.570805 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.570907 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.070888963 +0000 UTC m=+145.656189170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.612036 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.621312 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s9jgl" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.627780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vc6gw" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.630787 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.780136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.783820 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.283796373 +0000 UTC m=+145.869096580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.837474 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.846305 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.847759 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.871830 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z2ps9"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.871892 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.887161 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.887517 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.387502567 +0000 UTC m=+145.972802774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.894689 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm"] Jan 20 08:51:50 crc kubenswrapper[4967]: I0120 08:51:50.988755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:50 crc kubenswrapper[4967]: E0120 08:51:50.989136 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.489124411 +0000 UTC m=+146.074424618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.089426 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.089932 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.589916281 +0000 UTC m=+146.175216488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.099743 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.190471 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.190835 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.690825234 +0000 UTC m=+146.276125431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.291911 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.292521 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.792505939 +0000 UTC m=+146.377806146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.300504 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" event={"ID":"b3d6730d-549e-4a20-af1a-3096aab8c184","Type":"ContainerStarted","Data":"619808cb1b5c569c4680c30c29234ff4a333271d3744fff4b45e8863fe441acc"} Jan 20 08:51:51 crc kubenswrapper[4967]: W0120 08:51:51.300604 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c931549_0309_4d10_8908_7c972199f0df.slice/crio-24ce3bbcc63237e8d092fbb9932d9f578de98f15b3faa110af814d9071812a86 WatchSource:0}: Error finding container 24ce3bbcc63237e8d092fbb9932d9f578de98f15b3faa110af814d9071812a86: Status 404 returned error can't find the container with id 24ce3bbcc63237e8d092fbb9932d9f578de98f15b3faa110af814d9071812a86 Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.310287 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" event={"ID":"a3e9f372-2ab2-4caa-a9e2-4c563a446126","Type":"ContainerStarted","Data":"a13dd5242d79af1f5d15dc59f5c9bd0b12b05406d754ff6105f292899ca945fd"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.342599 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7pllc" event={"ID":"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589","Type":"ContainerStarted","Data":"d4b006369bddc30b5e0f410450b1a229fe246fbc36267886b5d2bb5972e9d6cd"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.364790 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vc6gw" event={"ID":"0173c2dd-2801-47b0-abb3-684fabcb1845","Type":"ContainerStarted","Data":"af20f100800269adff6e336405d05458b37bc4ccf5fb9726b615f8748046b483"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.369363 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z2ps9" event={"ID":"63125d8f-2be4-492c-8b42-057c5035715a","Type":"ContainerStarted","Data":"455c5be18cf32eaabbff366f9d5e9a6d87c6ec161d8c3e8a3894583e4524ddcb"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.389410 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" event={"ID":"f6f75790-6be7-4f4c-ae17-5258a715666e","Type":"ContainerStarted","Data":"fd99e397f075de2fb14f7b271a124b2f003d1da6f73fecf50e1676c7b7fa0dd3"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.393147 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.393391 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.893380032 +0000 UTC m=+146.478680239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.395448 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" event={"ID":"e05474a9-df95-4593-a3f5-ae1e8cb3785f","Type":"ContainerStarted","Data":"09c2d3ea49ddde7cf5e2b626d0f1e0402226297a852367d38085f710e9c1c256"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.454936 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" event={"ID":"8155bf3a-4152-4f9c-a9aa-7e3abcf84912","Type":"ContainerStarted","Data":"76d48fb5718a65d4cac3bf7349deccae29d4d32d3511aeddae03290467a9f5e5"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.459178 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" event={"ID":"c6512677-6466-4738-a2d8-726f0f85a9fd","Type":"ContainerStarted","Data":"e1991c68371b8b1d2bd250e973bbbadad724ed496ab1484f67ea734a5e46ae6b"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.479471 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" event={"ID":"ace51544-8fe2-4352-bcd8-93611e156b80","Type":"ContainerStarted","Data":"452dd45ebe51c97c83e7e497c01435a1074da738591d35f8860a1789fc54e1b7"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.498522 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.499295 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:51.9992694 +0000 UTC m=+146.584569607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.506097 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qfnnr" podStartSLOduration=127.506083129 podStartE2EDuration="2m7.506083129s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:51.50336854 +0000 UTC m=+146.088668747" watchObservedRunningTime="2026-01-20 08:51:51.506083129 +0000 UTC m=+146.091383336" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.529163 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" event={"ID":"ae790671-4e34-4dba-86ed-6d2851366835","Type":"ContainerStarted","Data":"d5aab0e8c18017f169f0485f3db9ee74c961d9a6f312b6916de330c53f66c01d"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.551820 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" event={"ID":"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b","Type":"ContainerStarted","Data":"75a02b7c6f17156b5bf087e1c74eea16186f454e43fd6df4dd34a4d990941fed"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.566226 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pm5cd" event={"ID":"adcafa80-4a5f-4268-8a01-6284b6cf235c","Type":"ContainerStarted","Data":"2b0e61077527e88e2a6800f762b67dca7e1c2574d0db86d56e475a18ad7a6eae"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.577697 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" event={"ID":"89ee25c2-17c2-4464-9dad-e9126c1fa0cf","Type":"ContainerStarted","Data":"59dbb54fc5d19a702f11d4e568d5afed981eba463dedf473a8d17b15fc216786"} Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.599455 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.599933 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.099917336 +0000 UTC m=+146.685217543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.675205 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.691149 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-75ccq"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.700020 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.700387 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.200347735 +0000 UTC m=+146.785647942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.800921 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.801148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.801168 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.801205 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.801239 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.801494 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.301481405 +0000 UTC m=+146.886781612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.802284 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.823047 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.827469 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.827874 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.859485 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" podStartSLOduration=126.859461456 podStartE2EDuration="2m6.859461456s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:51.85617614 +0000 UTC m=+146.441476367" watchObservedRunningTime="2026-01-20 08:51:51.859461456 +0000 UTC m=+146.444761663" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.867696 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xp9k6"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.886692 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.901753 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:51 crc kubenswrapper[4967]: E0120 08:51:51.902025 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.402009467 +0000 UTC m=+146.987309664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.910440 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.917129 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.918412 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.925551 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:51 crc kubenswrapper[4967]: I0120 08:51:51.934281 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.006959 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.007349 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.507333718 +0000 UTC m=+147.092633925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.056128 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8c5f152_a3c7_41b7_ae9d_2317eed21246.slice/crio-f92713e2557b8c1716fb705aabf988f0e19584494173e212b0d89f629b96451d WatchSource:0}: Error finding container f92713e2557b8c1716fb705aabf988f0e19584494173e212b0d89f629b96451d: Status 404 returned error can't find the container with id f92713e2557b8c1716fb705aabf988f0e19584494173e212b0d89f629b96451d Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.059518 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf0c805a_669e_422a_906a_917c335ed5ab.slice/crio-0b43f39750ad28af6d4d18fccde47df6f63cc905b76222f230bd3ac538fbb5ae WatchSource:0}: Error finding container 0b43f39750ad28af6d4d18fccde47df6f63cc905b76222f230bd3ac538fbb5ae: Status 404 returned error can't find the container with id 0b43f39750ad28af6d4d18fccde47df6f63cc905b76222f230bd3ac538fbb5ae Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.078806 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddefe6cb5_e637_4f50_a67a_7d1c6ae28879.slice/crio-d547ae833f1c3604779048bda3227dc881f73c5ca28611cfa0837ed751bffb83 WatchSource:0}: Error finding container d547ae833f1c3604779048bda3227dc881f73c5ca28611cfa0837ed751bffb83: Status 404 returned error can't find the container with id d547ae833f1c3604779048bda3227dc881f73c5ca28611cfa0837ed751bffb83 Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.107806 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.115906 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.615879944 +0000 UTC m=+147.201180151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.209763 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.210278 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.710267657 +0000 UTC m=+147.295567864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.313310 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.313801 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.813786807 +0000 UTC m=+147.399087014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.361042 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.377149 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.402306 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.404051 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.411525 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.413007 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.414918 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.415232 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:52.915221175 +0000 UTC m=+147.500521382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.417706 4967 csr.go:261] certificate signing request csr-fmhwm is approved, waiting to be issued Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.437123 4967 csr.go:257] certificate signing request csr-fmhwm is issued Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.503229 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podced462f3_f1bf_440f_bd5a_6973fc418386.slice/crio-8a41cd78d0c5293db418805d4276fe3269ddb43d88a1f63d3959d705ef9e93e7 WatchSource:0}: Error finding container 8a41cd78d0c5293db418805d4276fe3269ddb43d88a1f63d3959d705ef9e93e7: Status 404 returned error can't find the container with id 8a41cd78d0c5293db418805d4276fe3269ddb43d88a1f63d3959d705ef9e93e7 Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.515579 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.520073 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.020053963 +0000 UTC m=+147.605354170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.572220 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s9jgl"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.579929 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9fk7s"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.582213 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk"] Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.602349 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbc6eec0_bb7b_4d54_8cdb_21a43903e3f8.slice/crio-5615147e77f4ee8c4dadb956ec8ddbf935bfc2317db11e5ed145e596f395909d WatchSource:0}: Error finding container 5615147e77f4ee8c4dadb956ec8ddbf935bfc2317db11e5ed145e596f395909d: Status 404 returned error can't find the container with id 5615147e77f4ee8c4dadb956ec8ddbf935bfc2317db11e5ed145e596f395909d Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.617212 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.617520 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.117508735 +0000 UTC m=+147.702808942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.675734 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4fzjb"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.688059 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" event={"ID":"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7","Type":"ContainerStarted","Data":"ce4adff8de8859c0c88f862a811b52d99ad2a250fb4a7b8c2de372f4bb55b0b0"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.688918 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.700816 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-txfwl"] Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.715786 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" podStartSLOduration=128.715767951 podStartE2EDuration="2m8.715767951s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:52.714973418 +0000 UTC m=+147.300273625" watchObservedRunningTime="2026-01-20 08:51:52.715767951 +0000 UTC m=+147.301068158" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.718690 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.719128 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.219112969 +0000 UTC m=+147.804413176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.734766 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" event={"ID":"8c931549-0309-4d10-8908-7c972199f0df","Type":"ContainerStarted","Data":"6d56c66f1784d9a0040a057e3d1d75266dc7cc6af216f8cc14dd786b28b5101d"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.734807 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" event={"ID":"8c931549-0309-4d10-8908-7c972199f0df","Type":"ContainerStarted","Data":"24ce3bbcc63237e8d092fbb9932d9f578de98f15b3faa110af814d9071812a86"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.808302 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rlj75" podStartSLOduration=128.808283549 podStartE2EDuration="2m8.808283549s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:52.788133401 +0000 UTC m=+147.373433608" watchObservedRunningTime="2026-01-20 08:51:52.808283549 +0000 UTC m=+147.393583756" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.811380 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" event={"ID":"8db30b00-67b2-4713-bb9e-1b853cbf08e1","Type":"ContainerStarted","Data":"7889c148341a80c071df1c54dc698d61ea37a9e7d1724ffa87437eaa1bdae9f0"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.812076 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.819787 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.820150 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.320138565 +0000 UTC m=+147.905438772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.821153 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" event={"ID":"cf0c805a-669e-422a-906a-917c335ed5ab","Type":"ContainerStarted","Data":"0b43f39750ad28af6d4d18fccde47df6f63cc905b76222f230bd3ac538fbb5ae"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.826264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" event={"ID":"7db3ffb6-ef1c-432b-8235-5ef6d77525a4","Type":"ContainerStarted","Data":"9f8171ab6a0fe188c4f98a89800bf41a0a4d5b2edd9f048a0b7fa93cb63cd050"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.827549 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.838425 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" podStartSLOduration=128.838406747 podStartE2EDuration="2m8.838406747s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:52.83334361 +0000 UTC m=+147.418643817" watchObservedRunningTime="2026-01-20 08:51:52.838406747 +0000 UTC m=+147.423706954" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.865073 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-h6s8s" Jan 20 08:51:52 crc kubenswrapper[4967]: W0120 08:51:52.876178 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-98450a09eb8631326ef8008c37b01e24d601153e1dbf0b4cbfacaa7cf5371c4b WatchSource:0}: Error finding container 98450a09eb8631326ef8008c37b01e24d601153e1dbf0b4cbfacaa7cf5371c4b: Status 404 returned error can't find the container with id 98450a09eb8631326ef8008c37b01e24d601153e1dbf0b4cbfacaa7cf5371c4b Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.899929 4967 generic.go:334] "Generic (PLEG): container finished" podID="8155bf3a-4152-4f9c-a9aa-7e3abcf84912" containerID="76d48fb5718a65d4cac3bf7349deccae29d4d32d3511aeddae03290467a9f5e5" exitCode=0 Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.899999 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" event={"ID":"8155bf3a-4152-4f9c-a9aa-7e3abcf84912","Type":"ContainerDied","Data":"76d48fb5718a65d4cac3bf7349deccae29d4d32d3511aeddae03290467a9f5e5"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.900550 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.910535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7pllc" event={"ID":"a00ea83f-d3ff-4f4a-a82f-83eaf4edd589","Type":"ContainerStarted","Data":"4077c84020bbfc9c99a8404bd3110d98c469cf77376e0b144ee271a52b97f22e"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.921368 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:52 crc kubenswrapper[4967]: E0120 08:51:52.922633 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.422598283 +0000 UTC m=+148.007898490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.923387 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" event={"ID":"f6f75790-6be7-4f4c-ae17-5258a715666e","Type":"ContainerStarted","Data":"7573046a39c1d9f28b049b08ca21df7b79eab5513c332b14a8a752638bb65157"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.958041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" event={"ID":"cc8e9166-fdeb-4308-a44e-e932397bfead","Type":"ContainerStarted","Data":"d81e6a5ad7f248535eb26da71b0fc08f59a780684370dba7ae7688b01092921e"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.970133 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" event={"ID":"074019c5-93cc-496d-9ddd-49bac54ccfa5","Type":"ContainerStarted","Data":"967c51149609b35f07ac3eb2d3f23e19a71575d87d3624c247ab4c50bf6e663d"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.978516 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7pllc" podStartSLOduration=128.978500133 podStartE2EDuration="2m8.978500133s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:52.977763922 +0000 UTC m=+147.563064129" watchObservedRunningTime="2026-01-20 08:51:52.978500133 +0000 UTC m=+147.563800340" Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.982465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vc6gw" event={"ID":"0173c2dd-2801-47b0-abb3-684fabcb1845","Type":"ContainerStarted","Data":"9276ed311d4bec21e8d965c252c37d0ac350de41c59cedca2d88f10d3d8cf0b7"} Jan 20 08:51:52 crc kubenswrapper[4967]: I0120 08:51:52.986956 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" podStartSLOduration=128.986937869 podStartE2EDuration="2m8.986937869s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:52.941060582 +0000 UTC m=+147.526360789" watchObservedRunningTime="2026-01-20 08:51:52.986937869 +0000 UTC m=+147.572238076" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.024352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.027045 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.527031999 +0000 UTC m=+148.112332206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.057531 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" podStartSLOduration=129.057513828 podStartE2EDuration="2m9.057513828s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.055098638 +0000 UTC m=+147.640398845" watchObservedRunningTime="2026-01-20 08:51:53.057513828 +0000 UTC m=+147.642814035" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.113200 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" event={"ID":"e05474a9-df95-4593-a3f5-ae1e8cb3785f","Type":"ContainerStarted","Data":"5c8ce644367e5f7e32362026d0a2743b1e003bd02066ce65990e02afef7f6783"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.114574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" event={"ID":"ced462f3-f1bf-440f-bd5a-6973fc418386","Type":"ContainerStarted","Data":"8a41cd78d0c5293db418805d4276fe3269ddb43d88a1f63d3959d705ef9e93e7"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.127925 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" event={"ID":"340acfd5-f5f9-4db9-8d90-d72a1ccac052","Type":"ContainerStarted","Data":"99dbb1ae08e9dd904adf8b28636ddf76338e14ea92cab0328cefd3e3f9651ab3"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.137267 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.138033 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.638009916 +0000 UTC m=+148.223310123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.139024 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gpj5p" podStartSLOduration=128.139008715 podStartE2EDuration="2m8.139008715s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.085292808 +0000 UTC m=+147.670593015" watchObservedRunningTime="2026-01-20 08:51:53.139008715 +0000 UTC m=+147.724308922" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.139118 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" podStartSLOduration=129.139113429 podStartE2EDuration="2m9.139113429s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.135214504 +0000 UTC m=+147.720514711" watchObservedRunningTime="2026-01-20 08:51:53.139113429 +0000 UTC m=+147.724413636" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.139121 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" event={"ID":"80899a1a-ffa8-4daa-9f19-d80ca422eea9","Type":"ContainerStarted","Data":"42ffeff8a5989ce2120dfc5a2d2b57580887ff347461118dd237a0e2911a0c76"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.145819 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z2ps9" event={"ID":"63125d8f-2be4-492c-8b42-057c5035715a","Type":"ContainerStarted","Data":"d84a5f9d39945219211901512215d1b9cb8829fb569de3c78a2e92022935cbaa"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.146679 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.151719 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerStarted","Data":"d547ae833f1c3604779048bda3227dc881f73c5ca28611cfa0837ed751bffb83"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.152729 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.196679 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ksh2s container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.196740 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.197155 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-pm5cd" event={"ID":"adcafa80-4a5f-4268-8a01-6284b6cf235c","Type":"ContainerStarted","Data":"4a3248e212d0b3a66fa33bac2e722033250be5364b2a1a08cbdceeffcd2d7ebf"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.197365 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-z2ps9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.197421 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z2ps9" podUID="63125d8f-2be4-492c-8b42-057c5035715a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.248723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.250151 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.750137476 +0000 UTC m=+148.335437683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.265207 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" event={"ID":"90b763ed-98ef-4961-86bc-a624bf3af585","Type":"ContainerStarted","Data":"86aa72c54417f3a79e1abebb0ec104d734f2ccafcd9e40da02b60d2ca4d7cd87"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.265244 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" event={"ID":"90b763ed-98ef-4961-86bc-a624bf3af585","Type":"ContainerStarted","Data":"068ab4bf3bc8af07de781f2411d5e317b876f7e6c17f7b7ebaa0463d47346732"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.292366 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" podStartSLOduration=128.292344597 podStartE2EDuration="2m8.292344597s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.24650501 +0000 UTC m=+147.831805227" watchObservedRunningTime="2026-01-20 08:51:53.292344597 +0000 UTC m=+147.877644804" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.293500 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" podStartSLOduration=129.293493841 podStartE2EDuration="2m9.293493841s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.167148906 +0000 UTC m=+147.752449113" watchObservedRunningTime="2026-01-20 08:51:53.293493841 +0000 UTC m=+147.878794048" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.306699 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rm9cb" podStartSLOduration=130.306678965 podStartE2EDuration="2m10.306678965s" podCreationTimestamp="2026-01-20 08:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.297782726 +0000 UTC m=+147.883082943" watchObservedRunningTime="2026-01-20 08:51:53.306678965 +0000 UTC m=+147.891979182" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.326379 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" event={"ID":"334064e9-09d2-4cd7-8eb4-75449c214712","Type":"ContainerStarted","Data":"ce1a22e5659ae849678de2634d1fe1f71d89f3e8924993678e6b38179d6a2ccd"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.339451 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xp9k6" event={"ID":"b8c5f152-a3c7-41b7-ae9d-2317eed21246","Type":"ContainerStarted","Data":"f92713e2557b8c1716fb705aabf988f0e19584494173e212b0d89f629b96451d"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.365271 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.366236 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.866221593 +0000 UTC m=+148.451521800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.370028 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.398422 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vc6gw" podStartSLOduration=6.398399531 podStartE2EDuration="6.398399531s" podCreationTimestamp="2026-01-20 08:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.377996096 +0000 UTC m=+147.963296303" watchObservedRunningTime="2026-01-20 08:51:53.398399531 +0000 UTC m=+147.983699738" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.407673 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" event={"ID":"e4949fe9-aa28-4409-bf01-ed9247a422b3","Type":"ContainerStarted","Data":"d0dffc0c9ba5b35327788b048c3926218ee2bab2aafb284c74493012225c0519"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.435911 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" event={"ID":"2c82db03-f0fc-46b7-b71b-fb9ba14a2c7b","Type":"ContainerStarted","Data":"2ba7264a4eb8f3f6c6f892948c31a862ba6bcd05d05d5b04aa6163ebef88c0de"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.438990 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-20 08:46:52 +0000 UTC, rotation deadline is 2026-11-18 19:33:45.122380045 +0000 UTC Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.439031 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7258h41m51.683351419s for next certificate rotation Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.472421 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.473115 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:53.97310148 +0000 UTC m=+148.558401677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.483022 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" event={"ID":"89ee25c2-17c2-4464-9dad-e9126c1fa0cf","Type":"ContainerStarted","Data":"b6347d0febdb7808dd384ffcf072a5d0988aef2d14349dee6f63306ae9fa67dd"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.490823 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" event={"ID":"293c612b-323d-4408-af28-b9f559766583","Type":"ContainerStarted","Data":"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.491786 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.521366 4967 generic.go:334] "Generic (PLEG): container finished" podID="76b01181-c4c3-4ae6-8768-965aa2a235a3" containerID="7fa1bb734b9a8f5da3f8d6e65db312da4dc69c9a92268177d7c9a05e728a1ccb" exitCode=0 Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.521853 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-plddg" event={"ID":"76b01181-c4c3-4ae6-8768-965aa2a235a3","Type":"ContainerDied","Data":"7fa1bb734b9a8f5da3f8d6e65db312da4dc69c9a92268177d7c9a05e728a1ccb"} Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.522373 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.574158 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.574477 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.074452176 +0000 UTC m=+148.659752383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.574769 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.575061 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.075053993 +0000 UTC m=+148.660354200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.649083 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" podStartSLOduration=128.649064742 podStartE2EDuration="2m8.649064742s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.648651769 +0000 UTC m=+148.233951996" watchObservedRunningTime="2026-01-20 08:51:53.649064742 +0000 UTC m=+148.234364969" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.679224 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ssxc" podStartSLOduration=129.67920433 podStartE2EDuration="2m9.67920433s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.579851333 +0000 UTC m=+148.165151540" watchObservedRunningTime="2026-01-20 08:51:53.67920433 +0000 UTC m=+148.264504537" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.680835 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.682428 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.182408164 +0000 UTC m=+148.767708371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.688315 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-z2ps9" podStartSLOduration=129.688299046 podStartE2EDuration="2m9.688299046s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.687218634 +0000 UTC m=+148.272518841" watchObservedRunningTime="2026-01-20 08:51:53.688299046 +0000 UTC m=+148.273599253" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.725699 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-pm5cd" podStartSLOduration=129.725682066 podStartE2EDuration="2m9.725682066s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.723673698 +0000 UTC m=+148.308973905" watchObservedRunningTime="2026-01-20 08:51:53.725682066 +0000 UTC m=+148.310982273" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.754154 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" podStartSLOduration=129.754136126 podStartE2EDuration="2m9.754136126s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.75013257 +0000 UTC m=+148.335432797" watchObservedRunningTime="2026-01-20 08:51:53.754136126 +0000 UTC m=+148.339436333" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.789289 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.789668 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.289655482 +0000 UTC m=+148.874955689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.832292 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fz8kc" podStartSLOduration=129.832274205 podStartE2EDuration="2m9.832274205s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.796478321 +0000 UTC m=+148.381778528" watchObservedRunningTime="2026-01-20 08:51:53.832274205 +0000 UTC m=+148.417574412" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.891659 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:53 crc kubenswrapper[4967]: E0120 08:51:53.892104 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.39208969 +0000 UTC m=+148.977389897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.910126 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q7jxn" podStartSLOduration=129.910110586 podStartE2EDuration="2m9.910110586s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.833596733 +0000 UTC m=+148.418896940" watchObservedRunningTime="2026-01-20 08:51:53.910110586 +0000 UTC m=+148.495410793" Jan 20 08:51:53 crc kubenswrapper[4967]: I0120 08:51:53.951736 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" podStartSLOduration=128.951718609 podStartE2EDuration="2m8.951718609s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.911881087 +0000 UTC m=+148.497181294" watchObservedRunningTime="2026-01-20 08:51:53.951718609 +0000 UTC m=+148.537018816" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.005140 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.005514 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.505499158 +0000 UTC m=+149.090799365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.108380 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.108548 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.608528963 +0000 UTC m=+149.193829170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.109197 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.109496 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.609484971 +0000 UTC m=+149.194785178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.200998 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.215411 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.215802 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.715785291 +0000 UTC m=+149.301085498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.217700 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:54 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:54 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:54 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.217754 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.330738 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.331869 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.831851396 +0000 UTC m=+149.417151603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.437167 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.437590 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:54.93757632 +0000 UTC m=+149.522876527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.538971 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.551682 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.051663308 +0000 UTC m=+149.636963515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.600112 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xp9k6" event={"ID":"b8c5f152-a3c7-41b7-ae9d-2317eed21246","Type":"ContainerStarted","Data":"142e89fb6a630e43da0289a507f629500708326c094b288911617887ce1efa3d"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.600171 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xp9k6" event={"ID":"b8c5f152-a3c7-41b7-ae9d-2317eed21246","Type":"ContainerStarted","Data":"9c93e47185071b8bb7b0a5d16a317193f32184685c9dad6e53694e579b7085d2"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.600365 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xp9k6" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.601511 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xrwv5" event={"ID":"e8e4dde4-2a5b-4f66-95cf-ddfa13b8b1f7","Type":"ContainerStarted","Data":"8587aab77af0b32340eb1454a62f4cbe398b68238478e867b1cb31d0bd37151b"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.603148 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" event={"ID":"074019c5-93cc-496d-9ddd-49bac54ccfa5","Type":"ContainerStarted","Data":"67c16028367614b8ad8b903fc541f0b0ed63ba6fd56225e17b5419808784918c"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.603734 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.605721 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6e74d9643ac2eb3eca39a28f2b232e658fe33d58a7cccb078b0043fa20d00ef6"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.605748 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"44c2280d99e2460850c7f752e70e78f6e0632f5dba101d57d095dd78a73b5b80"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.613282 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerStarted","Data":"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.614485 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ksh2s container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.614521 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.625096 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" event={"ID":"5e360baa-93ae-4aa9-b2db-b362d45b9fed","Type":"ContainerStarted","Data":"b13455ab136c92f47f5fb9b918f8a5d6966aa02e3d89311c690e06a7b3a8f804"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.625151 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" event={"ID":"5e360baa-93ae-4aa9-b2db-b362d45b9fed","Type":"ContainerStarted","Data":"6fed1a6ed74842fef016924fcb0e40f85a79b4072b1779a1f2504b27caad1bbb"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.637037 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pm4ww" event={"ID":"ace51544-8fe2-4352-bcd8-93611e156b80","Type":"ContainerStarted","Data":"d897ea5a4760310f60c1b2901377377d90e44595e05558214fc8706026e2dc7f"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.653879 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" podStartSLOduration=130.653862178 podStartE2EDuration="2m10.653862178s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:53.988961165 +0000 UTC m=+148.574261372" watchObservedRunningTime="2026-01-20 08:51:54.653862178 +0000 UTC m=+149.239162385" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.657417 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.658439 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.158424171 +0000 UTC m=+149.743724378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.692904 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c9qgm" event={"ID":"c6512677-6466-4738-a2d8-726f0f85a9fd","Type":"ContainerStarted","Data":"8925b807a971f8cf4ca040fe9b70f229043accb1dc514b50e2503d8b6f12e634"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.704767 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" event={"ID":"334064e9-09d2-4cd7-8eb4-75449c214712","Type":"ContainerStarted","Data":"73533bd151607d051da43e6dacce101a24f0fd41e3a5417c5b73a2f2fc0cbe11"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.708969 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1f3c75ae9ce3947f43de5a5d644d1aa2ac7ac8b51dc6aad723a5f04e56cd9a82"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.709002 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b6f95d7d1eb5a5fa28e870a124466404cc4d23b5894f20c2505ff276f0c5f992"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.718527 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xp9k6" podStartSLOduration=7.718509824 podStartE2EDuration="7.718509824s" podCreationTimestamp="2026-01-20 08:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:54.663702025 +0000 UTC m=+149.249002232" watchObservedRunningTime="2026-01-20 08:51:54.718509824 +0000 UTC m=+149.303810031" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.750668 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" event={"ID":"55129ade-282b-431d-8e52-00e8396cfeaf","Type":"ContainerStarted","Data":"e7de6bae347e7c8b28a1d339a6d8903e1df6daaefc7483d99b551a1e12598b68"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.750702 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" event={"ID":"55129ade-282b-431d-8e52-00e8396cfeaf","Type":"ContainerStarted","Data":"da9707de21a65f97b9e6515302d514e78d62511e52e6151ffb14a6809fab6665"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.752097 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" podStartSLOduration=129.752087793 podStartE2EDuration="2m9.752087793s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:54.72079989 +0000 UTC m=+149.306100097" watchObservedRunningTime="2026-01-20 08:51:54.752087793 +0000 UTC m=+149.337388000" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.759232 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.761674 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.261661593 +0000 UTC m=+149.846961800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.830313 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" event={"ID":"ced462f3-f1bf-440f-bd5a-6973fc418386","Type":"ContainerStarted","Data":"5b02f42a660e62e38371a53b38ac247ff9b09af57aad3a699dc1ae15059aa830"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.830386 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.830399 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" event={"ID":"ced462f3-f1bf-440f-bd5a-6973fc418386","Type":"ContainerStarted","Data":"f04f20645951aecfd5cf9abe0a499e908baf765e64b16bacc379d22e232515c4"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.866671 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.867735 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.367717206 +0000 UTC m=+149.953017413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.870437 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" event={"ID":"a3e9f372-2ab2-4caa-a9e2-4c563a446126","Type":"ContainerStarted","Data":"566e4dea0b6b7c71700de3932047045835ac90317338f1ef63475e9ca6971a6f"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.870466 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" event={"ID":"a3e9f372-2ab2-4caa-a9e2-4c563a446126","Type":"ContainerStarted","Data":"c02ef103c4a32567c2e799d6d137282c313cf78d5fa4f0b6b6c923d536cb40fe"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.874908 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-75ccq" event={"ID":"e4949fe9-aa28-4409-bf01-ed9247a422b3","Type":"ContainerStarted","Data":"0acbe9e734bccd6f030a20d50e4b4156e5f64fad5962766a237df050369279ce"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.876419 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" event={"ID":"a0d60d70-87ff-45b9-b7e3-053172ce2fe5","Type":"ContainerStarted","Data":"0bb177a02f4afc869ca4117c209c8b098feaf71f49144eb3f80b7ad8297d997b"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.877409 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" event={"ID":"cf0c805a-669e-422a-906a-917c335ed5ab","Type":"ContainerStarted","Data":"d1bf009a8881e5f8b8e957a6078d6989e8fa11eb0869897ed2a2468f01431d15"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.878304 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.879628 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" event={"ID":"8155bf3a-4152-4f9c-a9aa-7e3abcf84912","Type":"ContainerStarted","Data":"e8f9d055ed3d6599392f5548036eb3220e7fb86e6ab0a696a1e0df22cc7c6256"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.888301 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a7fbb749cb18c895d3602e0eee71d917a83166ba6bb08a275d99898b3a9c347f"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.888344 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"98450a09eb8631326ef8008c37b01e24d601153e1dbf0b4cbfacaa7cf5371c4b"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.888980 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.898922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x9x4t" event={"ID":"e05474a9-df95-4593-a3f5-ae1e8cb3785f","Type":"ContainerStarted","Data":"b31090822ceb8eef6cee193d2696341da5619e85d2ebf23f8bb7b8bd285c7600"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.900717 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" event={"ID":"80899a1a-ffa8-4daa-9f19-d80ca422eea9","Type":"ContainerStarted","Data":"7923fda7f2a58b1bd26060244a2547ab80e0be0e6265492fc7538c49189ddfe6"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.900735 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" event={"ID":"80899a1a-ffa8-4daa-9f19-d80ca422eea9","Type":"ContainerStarted","Data":"90df1a566bedd9bd3cd85aa005e157a14d2d0ed64a5cefe21bbed7caaca73386"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.902034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5gqsg" event={"ID":"cc8e9166-fdeb-4308-a44e-e932397bfead","Type":"ContainerStarted","Data":"9f6782652051fd96ca427b0bed04b1fc4b979f7fb2ae49bef039d0ca8567e500"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.903530 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" event={"ID":"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8","Type":"ContainerStarted","Data":"f150db0afbce30d398bb214746d067651636e7b5ea316635aea5e29ae9d36e7b"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.903548 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" event={"ID":"cbc6eec0-bb7b-4d54-8cdb-21a43903e3f8","Type":"ContainerStarted","Data":"5615147e77f4ee8c4dadb956ec8ddbf935bfc2317db11e5ed145e596f395909d"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.904307 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.905298 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" event={"ID":"0a72e607-9834-431b-8095-193ae6452de3","Type":"ContainerStarted","Data":"5219db393e81cf198544fe85d4fd6e37c28f0d45343726c2c448478dbef4d3e6"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.905325 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" event={"ID":"0a72e607-9834-431b-8095-193ae6452de3","Type":"ContainerStarted","Data":"31e95aca101e22670698188fa278b009a5f3a620426e47b6a8dc2d987b7d649d"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.906661 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" event={"ID":"27f66629-6bff-4ce5-820c-8aebe63600cc","Type":"ContainerStarted","Data":"7da3bf6b7eced6110e67ebfae6454d5b417da08ea826533d49cea5f072c245ed"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.906681 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" event={"ID":"27f66629-6bff-4ce5-820c-8aebe63600cc","Type":"ContainerStarted","Data":"aa2fe3f16d8cb2aa1fb4c960d01c8ca92158c184f426f2b14d792ce211f8de17"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.907911 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" event={"ID":"b3d6730d-549e-4a20-af1a-3096aab8c184","Type":"ContainerStarted","Data":"b6cf909ae02fbd32327136d98ded9b390538256a3d9d860b0d4a3d499592635b"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.909081 4967 generic.go:334] "Generic (PLEG): container finished" podID="ae790671-4e34-4dba-86ed-6d2851366835" containerID="fbd4da0dffafbd0a50efccf7395fbc848cb5f23d7d608777e701bc34c1b494d6" exitCode=0 Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.909124 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" event={"ID":"ae790671-4e34-4dba-86ed-6d2851366835","Type":"ContainerDied","Data":"fbd4da0dffafbd0a50efccf7395fbc848cb5f23d7d608777e701bc34c1b494d6"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.922530 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4bdk" podStartSLOduration=129.922515144 podStartE2EDuration="2m9.922515144s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:54.920959299 +0000 UTC m=+149.506259516" watchObservedRunningTime="2026-01-20 08:51:54.922515144 +0000 UTC m=+149.507815351" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.928566 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" event={"ID":"e38c077e-5d16-4200-a9bf-2eb9756fd5e7","Type":"ContainerStarted","Data":"8f0c2cbdb7ed1ad5d3ef9c2999498876c2da125265aba60517e16c466b8ab595"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.928621 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" event={"ID":"e38c077e-5d16-4200-a9bf-2eb9756fd5e7","Type":"ContainerStarted","Data":"76964b89eb42a1785f510dbb920574e7a5abbce4127bd84f4451a1ed679be9ae"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.958905 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s9jgl" event={"ID":"03a5f3fa-e077-4dae-be02-0a5967a2a15a","Type":"ContainerStarted","Data":"12ade87a1b73f585c7427a6784744205259ac21df815d71f9da84bca498c779c"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.958957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s9jgl" event={"ID":"03a5f3fa-e077-4dae-be02-0a5967a2a15a","Type":"ContainerStarted","Data":"18533d2c4b97d81f6211822b2439fae87a1027df156c03ee39450385df61491e"} Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.959820 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-z2ps9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.959859 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z2ps9" podUID="63125d8f-2be4-492c-8b42-057c5035715a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.961504 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.963111 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h56fr" Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.968077 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:54 crc kubenswrapper[4967]: E0120 08:51:54.970128 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.470114322 +0000 UTC m=+150.055414529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:54 crc kubenswrapper[4967]: I0120 08:51:54.997699 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" podStartSLOduration=131.997684146 podStartE2EDuration="2m11.997684146s" podCreationTimestamp="2026-01-20 08:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:54.996244694 +0000 UTC m=+149.581544901" watchObservedRunningTime="2026-01-20 08:51:54.997684146 +0000 UTC m=+149.582984353" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.040449 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xn64x" podStartSLOduration=131.040435723 podStartE2EDuration="2m11.040435723s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.040028751 +0000 UTC m=+149.625328958" watchObservedRunningTime="2026-01-20 08:51:55.040435723 +0000 UTC m=+149.625735930" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.073067 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.074658 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.57463016 +0000 UTC m=+150.159930417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.175548 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.176015 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.675984607 +0000 UTC m=+150.261284814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.210661 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:55 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:55 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:55 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.210745 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.276869 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.277510 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.777480107 +0000 UTC m=+150.362780314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.297444 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.378311 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.378736 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.87872194 +0000 UTC m=+150.464022147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.403135 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cfrcf" podStartSLOduration=131.403114452 podStartE2EDuration="2m11.403114452s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.173421282 +0000 UTC m=+149.758721489" watchObservedRunningTime="2026-01-20 08:51:55.403114452 +0000 UTC m=+149.988414659" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.481249 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.481565 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:55.981547269 +0000 UTC m=+150.566847476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.552567 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgjxq" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.583208 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.583712 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.083684578 +0000 UTC m=+150.668984775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.589840 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" podStartSLOduration=130.589818897 podStartE2EDuration="2m10.589818897s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.413901747 +0000 UTC m=+149.999201954" watchObservedRunningTime="2026-01-20 08:51:55.589818897 +0000 UTC m=+150.175119104" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.590949 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hr6g" podStartSLOduration=131.59094216 podStartE2EDuration="2m11.59094216s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.588671334 +0000 UTC m=+150.173971551" watchObservedRunningTime="2026-01-20 08:51:55.59094216 +0000 UTC m=+150.176242367" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.683913 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.684105 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.184087177 +0000 UTC m=+150.769387384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.684162 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.684448 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.184441597 +0000 UTC m=+150.769741804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.699303 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f42q5" podStartSLOduration=130.699287729 podStartE2EDuration="2m10.699287729s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.698495166 +0000 UTC m=+150.283795373" watchObservedRunningTime="2026-01-20 08:51:55.699287729 +0000 UTC m=+150.284587936" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.700116 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tfmfv" podStartSLOduration=131.700109404 podStartE2EDuration="2m11.700109404s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.656846882 +0000 UTC m=+150.242147089" watchObservedRunningTime="2026-01-20 08:51:55.700109404 +0000 UTC m=+150.285409611" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.770747 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9fk7s" podStartSLOduration=130.770726174 podStartE2EDuration="2m10.770726174s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.730859581 +0000 UTC m=+150.316159808" watchObservedRunningTime="2026-01-20 08:51:55.770726174 +0000 UTC m=+150.356026391" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.784982 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.785373 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.28535536 +0000 UTC m=+150.870655577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.807330 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6p4jt" podStartSLOduration=130.807312301 podStartE2EDuration="2m10.807312301s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.794909959 +0000 UTC m=+150.380210166" watchObservedRunningTime="2026-01-20 08:51:55.807312301 +0000 UTC m=+150.392612508" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.822506 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s9jgl" podStartSLOduration=8.822491933 podStartE2EDuration="8.822491933s" podCreationTimestamp="2026-01-20 08:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:55.822128202 +0000 UTC m=+150.407428409" watchObservedRunningTime="2026-01-20 08:51:55.822491933 +0000 UTC m=+150.407792130" Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.886675 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.887226 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.38721478 +0000 UTC m=+150.972514977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.982069 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" event={"ID":"a0d60d70-87ff-45b9-b7e3-053172ce2fe5","Type":"ContainerStarted","Data":"677c8490058b4fb2f31662ea2f49d2b2da2cd6a4a105ca964480f69834470855"} Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.987854 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:55 crc kubenswrapper[4967]: E0120 08:51:55.988489 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.488470555 +0000 UTC m=+151.073770762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:55 crc kubenswrapper[4967]: I0120 08:51:55.998041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" event={"ID":"5e360baa-93ae-4aa9-b2db-b362d45b9fed","Type":"ContainerStarted","Data":"6349f7c73d09a295393ecc2ca92fe2342f0a63fae621f57219a0b74f631d56db"} Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.010369 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-plddg" event={"ID":"76b01181-c4c3-4ae6-8768-965aa2a235a3","Type":"ContainerStarted","Data":"357d448e17c6fa8e1822f9a478357fef55e3ac94bd592b389f19de046f29004c"} Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.010411 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-plddg" event={"ID":"76b01181-c4c3-4ae6-8768-965aa2a235a3","Type":"ContainerStarted","Data":"0957dcf708396437790d53e6d2458ef05a03b3b4ed83dd34d788faf779e253ca"} Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.024727 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" event={"ID":"ae790671-4e34-4dba-86ed-6d2851366835","Type":"ContainerStarted","Data":"df1675734b5feb30ad59f88edd3f5e891bb8f8f8539e03d9ab0caee6242ffd81"} Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.078708 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" event={"ID":"e38c077e-5d16-4200-a9bf-2eb9756fd5e7","Type":"ContainerStarted","Data":"2bd2f4af198ba35c897e0628f2b3db51c41a4784ad328f1305edd19eeb19d19a"} Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.080394 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-z2ps9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.080446 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z2ps9" podUID="63125d8f-2be4-492c-8b42-057c5035715a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.081780 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ksh2s container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.081832 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.095539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.097479 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.597462334 +0000 UTC m=+151.182762541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.203210 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.203736 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.703692381 +0000 UTC m=+151.288992588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.205047 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.205390 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.705382061 +0000 UTC m=+151.290682268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.221376 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k8jmc" podStartSLOduration=131.221352567 podStartE2EDuration="2m11.221352567s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:56.193453353 +0000 UTC m=+150.778753570" watchObservedRunningTime="2026-01-20 08:51:56.221352567 +0000 UTC m=+150.806652794" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.221555 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:56 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:56 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:56 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.221636 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.267527 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4fzjb" podStartSLOduration=131.267507003 podStartE2EDuration="2m11.267507003s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:56.265330139 +0000 UTC m=+150.850630366" watchObservedRunningTime="2026-01-20 08:51:56.267507003 +0000 UTC m=+150.852807210" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.307270 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.307702 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.807680765 +0000 UTC m=+151.392980972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.322487 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-plddg" podStartSLOduration=132.322463626 podStartE2EDuration="2m12.322463626s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:56.313294799 +0000 UTC m=+150.898595016" watchObservedRunningTime="2026-01-20 08:51:56.322463626 +0000 UTC m=+150.907763843" Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.413577 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.413983 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:56.913969535 +0000 UTC m=+151.499269742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.518058 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.518509 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.018490813 +0000 UTC m=+151.603791020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.620551 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.620970 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.120954691 +0000 UTC m=+151.706254898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.722175 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.722361 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.222337319 +0000 UTC m=+151.807637526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.722429 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.722785 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.222769011 +0000 UTC m=+151.808069218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.823566 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.823831 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.323806298 +0000 UTC m=+151.909106505 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.824089 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.824368 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.324356964 +0000 UTC m=+151.909657171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.925389 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.925572 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.425546425 +0000 UTC m=+152.010846632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:56 crc kubenswrapper[4967]: I0120 08:51:56.925823 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:56 crc kubenswrapper[4967]: E0120 08:51:56.926207 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.426191724 +0000 UTC m=+152.011491931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.026947 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.027167 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.527117777 +0000 UTC m=+152.112417984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.027231 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.027545 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.5275361 +0000 UTC m=+152.112836307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.087523 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" event={"ID":"a0d60d70-87ff-45b9-b7e3-053172ce2fe5","Type":"ContainerStarted","Data":"c46e91fcf580d34cf609d09d1a2508f86c7fcabc063625df2ded5855b3aa1e82"} Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.087620 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" event={"ID":"a0d60d70-87ff-45b9-b7e3-053172ce2fe5","Type":"ContainerStarted","Data":"d942c043b71a413cf33c9c16e21d0a53468b9dd6c748ff2590034a31d6aea200"} Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.128166 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.129008 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.628987009 +0000 UTC m=+152.214287216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.203200 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:57 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:57 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:57 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.203265 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.230276 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.230679 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.730659074 +0000 UTC m=+152.315959331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.279590 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" podStartSLOduration=132.27954911 podStartE2EDuration="2m12.27954911s" podCreationTimestamp="2026-01-20 08:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:56.437736228 +0000 UTC m=+151.023036435" watchObservedRunningTime="2026-01-20 08:51:57.27954911 +0000 UTC m=+151.864849317" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.282794 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.284118 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.286214 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.299907 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.331029 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.331199 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.831171476 +0000 UTC m=+152.416471683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.331241 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kwcf\" (UniqueName: \"kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.331275 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.331379 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.331468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.331728 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.831709182 +0000 UTC m=+152.417009389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.419708 4967 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.432972 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.433170 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.93314386 +0000 UTC m=+152.518444067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433223 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kwcf\" (UniqueName: \"kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433250 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433290 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433330 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.433706 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:57.933686107 +0000 UTC m=+152.518986314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433740 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.433798 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.455364 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kwcf\" (UniqueName: \"kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf\") pod \"community-operators-jbrxm\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.467429 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.468358 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.473158 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.484226 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.537843 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.538265 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:58.038236586 +0000 UTC m=+152.623536793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.538532 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.538574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.538660 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdp77\" (UniqueName: \"kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.538700 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.539088 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:58.03907995 +0000 UTC m=+152.624380157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.617935 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.640286 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.640583 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.640685 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdp77\" (UniqueName: \"kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.640722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.641177 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.641269 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-20 08:51:58.141251271 +0000 UTC m=+152.726551488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.641540 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.675964 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.683354 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.694665 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdp77\" (UniqueName: \"kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77\") pod \"certified-operators-qmfmw\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.720794 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.742262 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.742314 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7x8h\" (UniqueName: \"kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.742348 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.742364 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: E0120 08:51:57.742656 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-20 08:51:58.242642708 +0000 UTC m=+152.827942915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8g5fl" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.782889 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.841809 4967 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-20T08:51:57.419994267Z","Handler":null,"Name":""} Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.854199 4967 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.854299 4967 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.861496 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.861756 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7x8h\" (UniqueName: \"kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.861831 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.861847 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.862826 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.867250 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.870118 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.874768 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.876087 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.888446 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.889413 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7x8h\" (UniqueName: \"kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h\") pod \"community-operators-9fwf7\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.963152 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.963204 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.963230 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89thw\" (UniqueName: \"kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:57 crc kubenswrapper[4967]: I0120 08:51:57.963281 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.017777 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.017842 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.022205 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.023007 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.024745 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.027031 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.027208 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.042866 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.065480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.065885 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.065912 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.065952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.065983 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89thw\" (UniqueName: \"kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.066648 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.066803 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.071096 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.076202 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8g5fl\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.108304 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89thw\" (UniqueName: \"kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw\") pod \"certified-operators-58rr8\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.148022 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerStarted","Data":"042919757c6c32f23230cde2713f81b51c2846cb94fe329e40327c2bd522af21"} Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.172211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" event={"ID":"a0d60d70-87ff-45b9-b7e3-053172ce2fe5","Type":"ContainerStarted","Data":"e29d89a866a49730b833071c0f741887e1452fe3f36f4c779da184f19ccfa85e"} Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.173217 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.173253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.173732 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.194963 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.198850 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-txfwl" podStartSLOduration=11.198831533 podStartE2EDuration="11.198831533s" podCreationTimestamp="2026-01-20 08:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:58.193870588 +0000 UTC m=+152.779170795" watchObservedRunningTime="2026-01-20 08:51:58.198831533 +0000 UTC m=+152.784131740" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.200997 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:58 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:58 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:58 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.201055 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.201571 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.217343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.262873 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.365693 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.551748 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.590765 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.653884 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:51:58 crc kubenswrapper[4967]: I0120 08:51:58.916478 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 20 08:51:58 crc kubenswrapper[4967]: W0120 08:51:58.920301 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7380611b_fa55_410e_abb9_4669d4b8112f.slice/crio-675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0 WatchSource:0}: Error finding container 675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0: Status 404 returned error can't find the container with id 675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0 Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.178434 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" event={"ID":"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f","Type":"ContainerStarted","Data":"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.178471 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" event={"ID":"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f","Type":"ContainerStarted","Data":"fa2a077b630c362edff062ec1cbe9d5f8871077d0e06c0af1851f3c4e2265e34"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.179289 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.181176 4967 generic.go:334] "Generic (PLEG): container finished" podID="20939b87-fa8e-4128-be15-edbd83a634b6" containerID="106f0f70565bffb32765d09698883dd6c4e55ceb27e255f9fcc9ab01500bc6e6" exitCode=0 Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.181219 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerDied","Data":"106f0f70565bffb32765d09698883dd6c4e55ceb27e255f9fcc9ab01500bc6e6"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.181232 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerStarted","Data":"75b0b98f6e44493351bc7d52b49657cd6a33c0c0c4a81ed8ef54b13bf323219d"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.182519 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.185632 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7380611b-fa55-410e-abb9-4669d4b8112f","Type":"ContainerStarted","Data":"675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.186807 4967 generic.go:334] "Generic (PLEG): container finished" podID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerID="f504f912811c35675f7fb315da1e265e39c046bc7b90bd2d743d00e099ba2c98" exitCode=0 Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.186846 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerDied","Data":"f504f912811c35675f7fb315da1e265e39c046bc7b90bd2d743d00e099ba2c98"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.186860 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerStarted","Data":"ffe8e4b7782f1863c79b4962b00b43cd2b7ec7ccb472f940d3d29f93e0a2009a"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.188480 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerID="17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3" exitCode=0 Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.188569 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerDied","Data":"17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.201984 4967 generic.go:334] "Generic (PLEG): container finished" podID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerID="3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c" exitCode=0 Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.202097 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerDied","Data":"3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.202139 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerStarted","Data":"162c24c565b61a43160e5a0cd43eac9ef917b28adccaebc682d91d9eace78da6"} Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.204278 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:51:59 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:51:59 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:51:59 crc kubenswrapper[4967]: healthz check failed Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.204320 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.204342 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" podStartSLOduration=135.20432258 podStartE2EDuration="2m15.20432258s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:51:59.203127425 +0000 UTC m=+153.788427652" watchObservedRunningTime="2026-01-20 08:51:59.20432258 +0000 UTC m=+153.789622787" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.333151 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.333200 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.341829 4967 patch_prober.go:28] interesting pod/apiserver-76f77b778f-plddg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]log ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]etcd ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/generic-apiserver-start-informers ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/max-in-flight-filter ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 20 08:51:59 crc kubenswrapper[4967]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 20 08:51:59 crc kubenswrapper[4967]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/project.openshift.io-projectcache ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-startinformers ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 20 08:51:59 crc kubenswrapper[4967]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 20 08:51:59 crc kubenswrapper[4967]: livez check failed Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.341886 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-plddg" podUID="76b01181-c4c3-4ae6-8768-965aa2a235a3" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.471051 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.472627 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.475690 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.487174 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.600844 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.600914 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.601049 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl4zq\" (UniqueName: \"kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.702288 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.702839 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.703201 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl4zq\" (UniqueName: \"kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.703144 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.702787 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.705459 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.709998 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.710377 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.715908 4967 patch_prober.go:28] interesting pod/console-f9d7485db-7pllc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.715971 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7pllc" podUID="a00ea83f-d3ff-4f4a-a82f-83eaf4edd589" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.729320 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl4zq\" (UniqueName: \"kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq\") pod \"redhat-marketplace-d9b7s\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.773530 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.773647 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.780783 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.789724 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-z2ps9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.789770 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-z2ps9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.789778 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z2ps9" podUID="63125d8f-2be4-492c-8b42-057c5035715a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.789818 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-z2ps9" podUID="63125d8f-2be4-492c-8b42-057c5035715a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.792929 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.874717 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.876826 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:51:59 crc kubenswrapper[4967]: I0120 08:51:59.903410 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.008553 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.008950 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.008980 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jslhp\" (UniqueName: \"kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.109079 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.110561 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.110619 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jslhp\" (UniqueName: \"kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.110692 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.111349 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.111350 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: W0120 08:52:00.126335 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod551f69d7_2c12_4474_9169_4540933734e1.slice/crio-d910483add5080a8ec3f3557c7c322e06a18b724baa77a625390b4607aecff01 WatchSource:0}: Error finding container d910483add5080a8ec3f3557c7c322e06a18b724baa77a625390b4607aecff01: Status 404 returned error can't find the container with id d910483add5080a8ec3f3557c7c322e06a18b724baa77a625390b4607aecff01 Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.127858 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jslhp\" (UniqueName: \"kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp\") pod \"redhat-marketplace-9pxql\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.196003 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.199388 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:00 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:00 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:00 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.199432 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.204746 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.210761 4967 generic.go:334] "Generic (PLEG): container finished" podID="7380611b-fa55-410e-abb9-4669d4b8112f" containerID="6d88dc936094cb2e0e8fa33e8720d3244696dc4106f42792d016b2481ff28c67" exitCode=0 Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.210955 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7380611b-fa55-410e-abb9-4669d4b8112f","Type":"ContainerDied","Data":"6d88dc936094cb2e0e8fa33e8720d3244696dc4106f42792d016b2481ff28c67"} Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.214699 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerStarted","Data":"d910483add5080a8ec3f3557c7c322e06a18b724baa77a625390b4607aecff01"} Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.220196 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8lfg9" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.247000 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.467341 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.474897 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.479319 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.483927 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.522256 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.522369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.522461 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf42b\" (UniqueName: \"kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.625113 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.625177 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.625210 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf42b\" (UniqueName: \"kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.625603 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.625897 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.645590 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf42b\" (UniqueName: \"kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b\") pod \"redhat-operators-2lz9v\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.668973 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.750514 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.814147 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.880938 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.884510 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.890392 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.929869 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5x59\" (UniqueName: \"kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.929965 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:00 crc kubenswrapper[4967]: I0120 08:52:00.930007 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.031458 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5x59\" (UniqueName: \"kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.031538 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.031566 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.032322 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.032446 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.059714 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5x59\" (UniqueName: \"kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59\") pod \"redhat-operators-pn7jw\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.157605 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:52:01 crc kubenswrapper[4967]: W0120 08:52:01.164113 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod588d30c5_1a85_42cf_8b19_e12755604f91.slice/crio-15b2fe3c025d75936fb3ce7325ee2bc61a83857717c415b832a46b461a205511 WatchSource:0}: Error finding container 15b2fe3c025d75936fb3ce7325ee2bc61a83857717c415b832a46b461a205511: Status 404 returned error can't find the container with id 15b2fe3c025d75936fb3ce7325ee2bc61a83857717c415b832a46b461a205511 Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.200660 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:01 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:01 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:01 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.200714 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.233892 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerID="b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c" exitCode=0 Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.233963 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerDied","Data":"b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c"} Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.233995 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerStarted","Data":"823b1349e871568e622694b8fa70f0785151b474687aad2e72f5d00429c5c88d"} Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.236052 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerStarted","Data":"15b2fe3c025d75936fb3ce7325ee2bc61a83857717c415b832a46b461a205511"} Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.237665 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.241737 4967 generic.go:334] "Generic (PLEG): container finished" podID="334064e9-09d2-4cd7-8eb4-75449c214712" containerID="73533bd151607d051da43e6dacce101a24f0fd41e3a5417c5b73a2f2fc0cbe11" exitCode=0 Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.241792 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" event={"ID":"334064e9-09d2-4cd7-8eb4-75449c214712","Type":"ContainerDied","Data":"73533bd151607d051da43e6dacce101a24f0fd41e3a5417c5b73a2f2fc0cbe11"} Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.256180 4967 generic.go:334] "Generic (PLEG): container finished" podID="551f69d7-2c12-4474-9169-4540933734e1" containerID="38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e" exitCode=0 Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.256264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerDied","Data":"38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e"} Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.660895 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.738726 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.756071 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access\") pod \"7380611b-fa55-410e-abb9-4669d4b8112f\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.756166 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir\") pod \"7380611b-fa55-410e-abb9-4669d4b8112f\" (UID: \"7380611b-fa55-410e-abb9-4669d4b8112f\") " Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.756297 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7380611b-fa55-410e-abb9-4669d4b8112f" (UID: "7380611b-fa55-410e-abb9-4669d4b8112f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.756713 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7380611b-fa55-410e-abb9-4669d4b8112f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.788877 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7380611b-fa55-410e-abb9-4669d4b8112f" (UID: "7380611b-fa55-410e-abb9-4669d4b8112f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:52:01 crc kubenswrapper[4967]: W0120 08:52:01.788904 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b655a5d_4022_48f0_a80e_e790a9235b3c.slice/crio-f2f0670a05d483844edd96e9772965c58d9fd28322eb4fc719e08f72f490d1be WatchSource:0}: Error finding container f2f0670a05d483844edd96e9772965c58d9fd28322eb4fc719e08f72f490d1be: Status 404 returned error can't find the container with id f2f0670a05d483844edd96e9772965c58d9fd28322eb4fc719e08f72f490d1be Jan 20 08:52:01 crc kubenswrapper[4967]: I0120 08:52:01.858973 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380611b-fa55-410e-abb9-4669d4b8112f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.199891 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:02 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:02 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:02 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.200251 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.278801 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7380611b-fa55-410e-abb9-4669d4b8112f","Type":"ContainerDied","Data":"675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0"} Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.278859 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675a1b77b876941686be8f365ff58379a752f3696addc160b58adf75522e50a0" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.278944 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.282490 4967 generic.go:334] "Generic (PLEG): container finished" podID="588d30c5-1a85-42cf-8b19-e12755604f91" containerID="6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a" exitCode=0 Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.282601 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerDied","Data":"6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a"} Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.284553 4967 generic.go:334] "Generic (PLEG): container finished" podID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerID="aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9" exitCode=0 Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.284698 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerDied","Data":"aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9"} Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.284787 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerStarted","Data":"f2f0670a05d483844edd96e9772965c58d9fd28322eb4fc719e08f72f490d1be"} Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.298265 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xp9k6" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.617567 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.678183 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume\") pod \"334064e9-09d2-4cd7-8eb4-75449c214712\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.678631 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjzsv\" (UniqueName: \"kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv\") pod \"334064e9-09d2-4cd7-8eb4-75449c214712\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.678689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume\") pod \"334064e9-09d2-4cd7-8eb4-75449c214712\" (UID: \"334064e9-09d2-4cd7-8eb4-75449c214712\") " Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.681367 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume" (OuterVolumeSpecName: "config-volume") pod "334064e9-09d2-4cd7-8eb4-75449c214712" (UID: "334064e9-09d2-4cd7-8eb4-75449c214712"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.689584 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "334064e9-09d2-4cd7-8eb4-75449c214712" (UID: "334064e9-09d2-4cd7-8eb4-75449c214712"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.689723 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv" (OuterVolumeSpecName: "kube-api-access-kjzsv") pod "334064e9-09d2-4cd7-8eb4-75449c214712" (UID: "334064e9-09d2-4cd7-8eb4-75449c214712"). InnerVolumeSpecName "kube-api-access-kjzsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.780703 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/334064e9-09d2-4cd7-8eb4-75449c214712-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.780748 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjzsv\" (UniqueName: \"kubernetes.io/projected/334064e9-09d2-4cd7-8eb4-75449c214712-kube-api-access-kjzsv\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:02 crc kubenswrapper[4967]: I0120 08:52:02.780762 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/334064e9-09d2-4cd7-8eb4-75449c214712-config-volume\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:03 crc kubenswrapper[4967]: I0120 08:52:03.200144 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:03 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:03 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:03 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:03 crc kubenswrapper[4967]: I0120 08:52:03.200414 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:03 crc kubenswrapper[4967]: I0120 08:52:03.297885 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" event={"ID":"334064e9-09d2-4cd7-8eb4-75449c214712","Type":"ContainerDied","Data":"ce1a22e5659ae849678de2634d1fe1f71d89f3e8924993678e6b38179d6a2ccd"} Jan 20 08:52:03 crc kubenswrapper[4967]: I0120 08:52:03.298373 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce1a22e5659ae849678de2634d1fe1f71d89f3e8924993678e6b38179d6a2ccd" Jan 20 08:52:03 crc kubenswrapper[4967]: I0120 08:52:03.298471 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v" Jan 20 08:52:04 crc kubenswrapper[4967]: I0120 08:52:04.198251 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:04 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:04 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:04 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:04 crc kubenswrapper[4967]: I0120 08:52:04.198306 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:04 crc kubenswrapper[4967]: I0120 08:52:04.339466 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:52:04 crc kubenswrapper[4967]: I0120 08:52:04.345492 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-plddg" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.197318 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:05 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:05 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:05 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.197676 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.494863 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 20 08:52:05 crc kubenswrapper[4967]: E0120 08:52:05.495090 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7380611b-fa55-410e-abb9-4669d4b8112f" containerName="pruner" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.495102 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7380611b-fa55-410e-abb9-4669d4b8112f" containerName="pruner" Jan 20 08:52:05 crc kubenswrapper[4967]: E0120 08:52:05.495121 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="334064e9-09d2-4cd7-8eb4-75449c214712" containerName="collect-profiles" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.495128 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="334064e9-09d2-4cd7-8eb4-75449c214712" containerName="collect-profiles" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.495226 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="334064e9-09d2-4cd7-8eb4-75449c214712" containerName="collect-profiles" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.495236 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7380611b-fa55-410e-abb9-4669d4b8112f" containerName="pruner" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.497021 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.504078 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.505000 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.505226 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.524192 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.524255 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.625855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.626008 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.626107 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.646404 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:05 crc kubenswrapper[4967]: I0120 08:52:05.828552 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.199984 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:06 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:06 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:06 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.200278 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.340523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.361455 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebe5f8d7-df5d-46aa-b33e-c3d6768bb245-metrics-certs\") pod \"network-metrics-daemon-lxp6p\" (UID: \"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245\") " pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.406890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lxp6p" Jan 20 08:52:06 crc kubenswrapper[4967]: I0120 08:52:06.476526 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 20 08:52:07 crc kubenswrapper[4967]: I0120 08:52:07.197594 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:07 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:07 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:07 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:07 crc kubenswrapper[4967]: I0120 08:52:07.197852 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:08 crc kubenswrapper[4967]: I0120 08:52:08.210993 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:08 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:08 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:08 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:08 crc kubenswrapper[4967]: I0120 08:52:08.211058 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:09 crc kubenswrapper[4967]: I0120 08:52:09.197226 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:09 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:09 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:09 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:09 crc kubenswrapper[4967]: I0120 08:52:09.197649 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:09 crc kubenswrapper[4967]: I0120 08:52:09.710987 4967 patch_prober.go:28] interesting pod/console-f9d7485db-7pllc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 20 08:52:09 crc kubenswrapper[4967]: I0120 08:52:09.711051 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7pllc" podUID="a00ea83f-d3ff-4f4a-a82f-83eaf4edd589" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 20 08:52:09 crc kubenswrapper[4967]: I0120 08:52:09.794573 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-z2ps9" Jan 20 08:52:10 crc kubenswrapper[4967]: I0120 08:52:10.199289 4967 patch_prober.go:28] interesting pod/router-default-5444994796-pm5cd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 20 08:52:10 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Jan 20 08:52:10 crc kubenswrapper[4967]: [+]process-running ok Jan 20 08:52:10 crc kubenswrapper[4967]: healthz check failed Jan 20 08:52:10 crc kubenswrapper[4967]: I0120 08:52:10.199375 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-pm5cd" podUID="adcafa80-4a5f-4268-8a01-6284b6cf235c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 20 08:52:11 crc kubenswrapper[4967]: I0120 08:52:11.196891 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:52:11 crc kubenswrapper[4967]: I0120 08:52:11.200729 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-pm5cd" Jan 20 08:52:17 crc kubenswrapper[4967]: W0120 08:52:17.857553 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod55b38ddc_1951_4356_8b18_f6d32a883fd0.slice/crio-712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e WatchSource:0}: Error finding container 712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e: Status 404 returned error can't find the container with id 712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e Jan 20 08:52:18 crc kubenswrapper[4967]: I0120 08:52:18.223679 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:52:18 crc kubenswrapper[4967]: I0120 08:52:18.473889 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:52:18 crc kubenswrapper[4967]: I0120 08:52:18.473952 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:52:18 crc kubenswrapper[4967]: I0120 08:52:18.475465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"55b38ddc-1951-4356-8b18-f6d32a883fd0","Type":"ContainerStarted","Data":"712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e"} Jan 20 08:52:19 crc kubenswrapper[4967]: I0120 08:52:19.715676 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:52:19 crc kubenswrapper[4967]: I0120 08:52:19.723194 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7pllc" Jan 20 08:52:30 crc kubenswrapper[4967]: I0120 08:52:30.524440 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pkrmf" Jan 20 08:52:32 crc kubenswrapper[4967]: I0120 08:52:32.073087 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 20 08:52:32 crc kubenswrapper[4967]: E0120 08:52:32.907944 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 20 08:52:32 crc kubenswrapper[4967]: E0120 08:52:32.908425 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mdp77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qmfmw_openshift-marketplace(5216e81e-0c81-4fc3-9fa2-458e24ed883d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:32 crc kubenswrapper[4967]: E0120 08:52:32.909724 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qmfmw" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" Jan 20 08:52:34 crc kubenswrapper[4967]: E0120 08:52:34.077006 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qmfmw" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" Jan 20 08:52:34 crc kubenswrapper[4967]: E0120 08:52:34.358589 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 20 08:52:34 crc kubenswrapper[4967]: E0120 08:52:34.359063 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-89thw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-58rr8_openshift-marketplace(1171240e-d46a-4eca-b9dc-fcc9e14b27a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:34 crc kubenswrapper[4967]: E0120 08:52:34.360275 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-58rr8" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" Jan 20 08:52:35 crc kubenswrapper[4967]: E0120 08:52:35.627982 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-58rr8" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" Jan 20 08:52:35 crc kubenswrapper[4967]: E0120 08:52:35.692627 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 20 08:52:35 crc kubenswrapper[4967]: E0120 08:52:35.692803 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n7x8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9fwf7_openshift-marketplace(20939b87-fa8e-4128-be15-edbd83a634b6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:35 crc kubenswrapper[4967]: E0120 08:52:35.694806 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9fwf7" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.317007 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.329219 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.329325 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.440761 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.440848 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.542176 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.542468 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.542590 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.567377 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:38 crc kubenswrapper[4967]: I0120 08:52:38.655004 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.006092 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9fwf7" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.098826 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.099007 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rf42b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2lz9v_openshift-marketplace(588d30c5-1a85-42cf-8b19-e12755604f91): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.100225 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2lz9v" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.104800 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.105332 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m5x59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pn7jw_openshift-marketplace(8b655a5d-4022-48f0-a80e-e790a9235b3c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.106497 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pn7jw" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.106938 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.107046 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8kwcf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jbrxm_openshift-marketplace(a0eaef46-8b17-45f6-9786-da4ebff48dfe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:39 crc kubenswrapper[4967]: E0120 08:52:39.108249 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jbrxm" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" Jan 20 08:52:42 crc kubenswrapper[4967]: E0120 08:52:42.636007 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pn7jw" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" Jan 20 08:52:42 crc kubenswrapper[4967]: E0120 08:52:42.636040 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2lz9v" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" Jan 20 08:52:42 crc kubenswrapper[4967]: E0120 08:52:42.636054 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jbrxm" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" Jan 20 08:52:42 crc kubenswrapper[4967]: I0120 08:52:42.897654 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 20 08:52:42 crc kubenswrapper[4967]: I0120 08:52:42.898913 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:42 crc kubenswrapper[4967]: I0120 08:52:42.911545 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.010775 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.010821 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.010845 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.032781 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lxp6p"] Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.111553 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.111695 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.111716 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.111734 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.111904 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.130267 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access\") pod \"installer-9-crc\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: I0120 08:52:43.227587 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:52:43 crc kubenswrapper[4967]: W0120 08:52:43.864522 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebe5f8d7_df5d_46aa_b33e_c3d6768bb245.slice/crio-77de36b11fb528997ca8d22d1acfdf8f41c43cd2982c3d37121b6624a577b376 WatchSource:0}: Error finding container 77de36b11fb528997ca8d22d1acfdf8f41c43cd2982c3d37121b6624a577b376: Status 404 returned error can't find the container with id 77de36b11fb528997ca8d22d1acfdf8f41c43cd2982c3d37121b6624a577b376 Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.952576 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.952941 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wl4zq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-d9b7s_openshift-marketplace(551f69d7-2c12-4474-9169-4540933734e1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.954379 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-d9b7s" podUID="551f69d7-2c12-4474-9169-4540933734e1" Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.962346 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.962541 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jslhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9pxql_openshift-marketplace(f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 20 08:52:43 crc kubenswrapper[4967]: E0120 08:52:43.965793 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9pxql" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.249773 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 20 08:52:44 crc kubenswrapper[4967]: W0120 08:52:44.260340 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod31c574c9_8911_4502_bd88_d3c530e26201.slice/crio-ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f WatchSource:0}: Error finding container ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f: Status 404 returned error can't find the container with id ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.318807 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 20 08:52:44 crc kubenswrapper[4967]: W0120 08:52:44.329773 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2199d68b_78b9_4479_b622_8c5d7fc2f1c8.slice/crio-16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c WatchSource:0}: Error finding container 16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c: Status 404 returned error can't find the container with id 16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.632136 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2199d68b-78b9-4479-b622-8c5d7fc2f1c8","Type":"ContainerStarted","Data":"16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.634869 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"31c574c9-8911-4502-bd88-d3c530e26201","Type":"ContainerStarted","Data":"cce11dbe2d82200879a79af6ff069cce3300b8df95ef027584af9e6692c4d9db"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.634900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"31c574c9-8911-4502-bd88-d3c530e26201","Type":"ContainerStarted","Data":"ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.636152 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"55b38ddc-1951-4356-8b18-f6d32a883fd0","Type":"ContainerStarted","Data":"35fe98dcc7b45460075c9ce7f681099d35af944c06a610901e425b0620b9cdcc"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.637853 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" event={"ID":"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245","Type":"ContainerStarted","Data":"a980ff9c03e57dd5a367119f36198ee45ef4a67724f319ba0b2da3ac03304b12"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.637877 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" event={"ID":"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245","Type":"ContainerStarted","Data":"4855109e8504982d93905f897afb8b500a613df6d3281b69ad1964753d06a4e9"} Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.637887 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lxp6p" event={"ID":"ebe5f8d7-df5d-46aa-b33e-c3d6768bb245","Type":"ContainerStarted","Data":"77de36b11fb528997ca8d22d1acfdf8f41c43cd2982c3d37121b6624a577b376"} Jan 20 08:52:44 crc kubenswrapper[4967]: E0120 08:52:44.639739 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9pxql" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" Jan 20 08:52:44 crc kubenswrapper[4967]: E0120 08:52:44.640016 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-d9b7s" podUID="551f69d7-2c12-4474-9169-4540933734e1" Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.652407 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=39.652388618 podStartE2EDuration="39.652388618s" podCreationTimestamp="2026-01-20 08:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:52:44.650550842 +0000 UTC m=+199.235851069" watchObservedRunningTime="2026-01-20 08:52:44.652388618 +0000 UTC m=+199.237688825" Jan 20 08:52:44 crc kubenswrapper[4967]: I0120 08:52:44.711230 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lxp6p" podStartSLOduration=180.711212644 podStartE2EDuration="3m0.711212644s" podCreationTimestamp="2026-01-20 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:52:44.709956343 +0000 UTC m=+199.295256570" watchObservedRunningTime="2026-01-20 08:52:44.711212644 +0000 UTC m=+199.296512851" Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.645179 4967 generic.go:334] "Generic (PLEG): container finished" podID="31c574c9-8911-4502-bd88-d3c530e26201" containerID="cce11dbe2d82200879a79af6ff069cce3300b8df95ef027584af9e6692c4d9db" exitCode=0 Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.645262 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"31c574c9-8911-4502-bd88-d3c530e26201","Type":"ContainerDied","Data":"cce11dbe2d82200879a79af6ff069cce3300b8df95ef027584af9e6692c4d9db"} Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.648458 4967 generic.go:334] "Generic (PLEG): container finished" podID="55b38ddc-1951-4356-8b18-f6d32a883fd0" containerID="35fe98dcc7b45460075c9ce7f681099d35af944c06a610901e425b0620b9cdcc" exitCode=0 Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.648523 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"55b38ddc-1951-4356-8b18-f6d32a883fd0","Type":"ContainerDied","Data":"35fe98dcc7b45460075c9ce7f681099d35af944c06a610901e425b0620b9cdcc"} Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.651885 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2199d68b-78b9-4479-b622-8c5d7fc2f1c8","Type":"ContainerStarted","Data":"a58c523040356958f654c1c6799a72ca5565e08e5fb5cffdba2d3c4e3e42dd5b"} Jan 20 08:52:45 crc kubenswrapper[4967]: I0120 08:52:45.707768 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.707747086 podStartE2EDuration="3.707747086s" podCreationTimestamp="2026-01-20 08:52:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:52:45.703496011 +0000 UTC m=+200.288796218" watchObservedRunningTime="2026-01-20 08:52:45.707747086 +0000 UTC m=+200.293047293" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.014522 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.020173 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073249 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access\") pod \"31c574c9-8911-4502-bd88-d3c530e26201\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073315 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access\") pod \"55b38ddc-1951-4356-8b18-f6d32a883fd0\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073352 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir\") pod \"31c574c9-8911-4502-bd88-d3c530e26201\" (UID: \"31c574c9-8911-4502-bd88-d3c530e26201\") " Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073421 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir\") pod \"55b38ddc-1951-4356-8b18-f6d32a883fd0\" (UID: \"55b38ddc-1951-4356-8b18-f6d32a883fd0\") " Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073707 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "31c574c9-8911-4502-bd88-d3c530e26201" (UID: "31c574c9-8911-4502-bd88-d3c530e26201"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.073721 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "55b38ddc-1951-4356-8b18-f6d32a883fd0" (UID: "55b38ddc-1951-4356-8b18-f6d32a883fd0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.078390 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "31c574c9-8911-4502-bd88-d3c530e26201" (UID: "31c574c9-8911-4502-bd88-d3c530e26201"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.078673 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "55b38ddc-1951-4356-8b18-f6d32a883fd0" (UID: "55b38ddc-1951-4356-8b18-f6d32a883fd0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.178430 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55b38ddc-1951-4356-8b18-f6d32a883fd0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.178467 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31c574c9-8911-4502-bd88-d3c530e26201-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.178481 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55b38ddc-1951-4356-8b18-f6d32a883fd0-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.178493 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31c574c9-8911-4502-bd88-d3c530e26201-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.664301 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"31c574c9-8911-4502-bd88-d3c530e26201","Type":"ContainerDied","Data":"ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f"} Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.664521 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab1ce6cbf221ef0707b6b8da8c03fcd6f15c93434a05455782eeae384a8c582f" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.664348 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.665550 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"55b38ddc-1951-4356-8b18-f6d32a883fd0","Type":"ContainerDied","Data":"712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e"} Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.665580 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 20 08:52:47 crc kubenswrapper[4967]: I0120 08:52:47.665582 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="712c5b682ccb0e96c69fbccf5441fddc2fdfe0fb2e9cfb07f460f55f9a3e685e" Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.474007 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.474078 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.474483 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.475136 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.475251 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b" gracePeriod=600 Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.673787 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b" exitCode=0 Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.673880 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b"} Jan 20 08:52:48 crc kubenswrapper[4967]: I0120 08:52:48.676408 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerStarted","Data":"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a"} Jan 20 08:52:49 crc kubenswrapper[4967]: I0120 08:52:49.682542 4967 generic.go:334] "Generic (PLEG): container finished" podID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerID="a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a" exitCode=0 Jan 20 08:52:49 crc kubenswrapper[4967]: I0120 08:52:49.682657 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerDied","Data":"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a"} Jan 20 08:52:49 crc kubenswrapper[4967]: I0120 08:52:49.686288 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6"} Jan 20 08:52:51 crc kubenswrapper[4967]: I0120 08:52:51.736326 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerStarted","Data":"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0"} Jan 20 08:52:51 crc kubenswrapper[4967]: I0120 08:52:51.756271 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qmfmw" podStartSLOduration=3.215312052 podStartE2EDuration="54.75625209s" podCreationTimestamp="2026-01-20 08:51:57 +0000 UTC" firstStartedPulling="2026-01-20 08:51:59.209885532 +0000 UTC m=+153.795185739" lastFinishedPulling="2026-01-20 08:52:50.75082557 +0000 UTC m=+205.336125777" observedRunningTime="2026-01-20 08:52:51.750261923 +0000 UTC m=+206.335562130" watchObservedRunningTime="2026-01-20 08:52:51.75625209 +0000 UTC m=+206.341552297" Jan 20 08:52:52 crc kubenswrapper[4967]: I0120 08:52:52.742671 4967 generic.go:334] "Generic (PLEG): container finished" podID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerID="ef2500f511c5ebecf19d7d823fd4988223d863a09186bb490c81f18f7be52c6c" exitCode=0 Jan 20 08:52:52 crc kubenswrapper[4967]: I0120 08:52:52.742769 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerDied","Data":"ef2500f511c5ebecf19d7d823fd4988223d863a09186bb490c81f18f7be52c6c"} Jan 20 08:52:53 crc kubenswrapper[4967]: I0120 08:52:53.750047 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerStarted","Data":"9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2"} Jan 20 08:52:53 crc kubenswrapper[4967]: I0120 08:52:53.751699 4967 generic.go:334] "Generic (PLEG): container finished" podID="20939b87-fa8e-4128-be15-edbd83a634b6" containerID="8dc8aa7c49e8862a71bef2e6cb613a4697210c9192d37dc9dcf4f516f0314e9b" exitCode=0 Jan 20 08:52:53 crc kubenswrapper[4967]: I0120 08:52:53.751731 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerDied","Data":"8dc8aa7c49e8862a71bef2e6cb613a4697210c9192d37dc9dcf4f516f0314e9b"} Jan 20 08:52:53 crc kubenswrapper[4967]: I0120 08:52:53.793543 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-58rr8" podStartSLOduration=2.572135444 podStartE2EDuration="56.79349826s" podCreationTimestamp="2026-01-20 08:51:57 +0000 UTC" firstStartedPulling="2026-01-20 08:51:59.18785351 +0000 UTC m=+153.773153717" lastFinishedPulling="2026-01-20 08:52:53.409216336 +0000 UTC m=+207.994516533" observedRunningTime="2026-01-20 08:52:53.768948616 +0000 UTC m=+208.354248823" watchObservedRunningTime="2026-01-20 08:52:53.79349826 +0000 UTC m=+208.378798467" Jan 20 08:52:54 crc kubenswrapper[4967]: I0120 08:52:54.767218 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerStarted","Data":"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b"} Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.777086 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerStarted","Data":"ed13ef5eeb4ee4987cde6255c3e782c34c9f1dca4777a6a83dcc49d06839d6cc"} Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.779603 4967 generic.go:334] "Generic (PLEG): container finished" podID="588d30c5-1a85-42cf-8b19-e12755604f91" containerID="83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b" exitCode=0 Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.779715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerDied","Data":"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b"} Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.783030 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerID="c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e" exitCode=0 Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.783071 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerDied","Data":"c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e"} Jan 20 08:52:55 crc kubenswrapper[4967]: I0120 08:52:55.824622 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9fwf7" podStartSLOduration=3.299066335 podStartE2EDuration="58.824588108s" podCreationTimestamp="2026-01-20 08:51:57 +0000 UTC" firstStartedPulling="2026-01-20 08:51:59.182315508 +0000 UTC m=+153.767615715" lastFinishedPulling="2026-01-20 08:52:54.707837271 +0000 UTC m=+209.293137488" observedRunningTime="2026-01-20 08:52:55.799943232 +0000 UTC m=+210.385243439" watchObservedRunningTime="2026-01-20 08:52:55.824588108 +0000 UTC m=+210.409888315" Jan 20 08:52:57 crc kubenswrapper[4967]: I0120 08:52:57.783450 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:52:57 crc kubenswrapper[4967]: I0120 08:52:57.783837 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:52:57 crc kubenswrapper[4967]: I0120 08:52:57.955459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerStarted","Data":"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416"} Jan 20 08:52:57 crc kubenswrapper[4967]: I0120 08:52:57.958604 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerStarted","Data":"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205"} Jan 20 08:52:57 crc kubenswrapper[4967]: I0120 08:52:57.974079 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2lz9v" podStartSLOduration=2.998834805 podStartE2EDuration="57.974064116s" podCreationTimestamp="2026-01-20 08:52:00 +0000 UTC" firstStartedPulling="2026-01-20 08:52:02.284418466 +0000 UTC m=+156.869718673" lastFinishedPulling="2026-01-20 08:52:57.259647777 +0000 UTC m=+211.844947984" observedRunningTime="2026-01-20 08:52:57.970792175 +0000 UTC m=+212.556092392" watchObservedRunningTime="2026-01-20 08:52:57.974064116 +0000 UTC m=+212.559364323" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.043190 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.043255 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.202544 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.202623 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.339902 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.340991 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.348309 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.360340 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jbrxm" podStartSLOduration=3.43348458 podStartE2EDuration="1m1.360325339s" podCreationTimestamp="2026-01-20 08:51:57 +0000 UTC" firstStartedPulling="2026-01-20 08:51:59.191460995 +0000 UTC m=+153.776761202" lastFinishedPulling="2026-01-20 08:52:57.118301754 +0000 UTC m=+211.703601961" observedRunningTime="2026-01-20 08:52:57.986529783 +0000 UTC m=+212.571830000" watchObservedRunningTime="2026-01-20 08:52:58.360325339 +0000 UTC m=+212.945625536" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.393020 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.965056 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerStarted","Data":"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6"} Jan 20 08:52:58 crc kubenswrapper[4967]: I0120 08:52:58.968452 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerStarted","Data":"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4"} Jan 20 08:52:59 crc kubenswrapper[4967]: I0120 08:52:59.051277 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:52:59 crc kubenswrapper[4967]: I0120 08:52:59.992863 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerID="20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4" exitCode=0 Jan 20 08:52:59 crc kubenswrapper[4967]: I0120 08:52:59.992946 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerDied","Data":"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4"} Jan 20 08:53:00 crc kubenswrapper[4967]: I0120 08:53:00.814813 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:53:00 crc kubenswrapper[4967]: I0120 08:53:00.814871 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:53:01 crc kubenswrapper[4967]: I0120 08:53:01.738799 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:53:01 crc kubenswrapper[4967]: I0120 08:53:01.857305 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2lz9v" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="registry-server" probeResult="failure" output=< Jan 20 08:53:01 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 08:53:01 crc kubenswrapper[4967]: > Jan 20 08:53:02 crc kubenswrapper[4967]: I0120 08:53:02.015442 4967 generic.go:334] "Generic (PLEG): container finished" podID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerID="55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6" exitCode=0 Jan 20 08:53:02 crc kubenswrapper[4967]: I0120 08:53:02.015504 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerDied","Data":"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6"} Jan 20 08:53:02 crc kubenswrapper[4967]: I0120 08:53:02.015827 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-58rr8" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="registry-server" containerID="cri-o://9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" gracePeriod=2 Jan 20 08:53:05 crc kubenswrapper[4967]: I0120 08:53:05.033436 4967 generic.go:334] "Generic (PLEG): container finished" podID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerID="9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" exitCode=0 Jan 20 08:53:05 crc kubenswrapper[4967]: I0120 08:53:05.033630 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerDied","Data":"9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2"} Jan 20 08:53:07 crc kubenswrapper[4967]: I0120 08:53:07.619514 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:53:07 crc kubenswrapper[4967]: I0120 08:53:07.619852 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:53:07 crc kubenswrapper[4967]: I0120 08:53:07.679501 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:53:08 crc kubenswrapper[4967]: I0120 08:53:08.085825 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:53:08 crc kubenswrapper[4967]: E0120 08:53:08.203367 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2 is running failed: container process not found" containerID="9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:53:08 crc kubenswrapper[4967]: E0120 08:53:08.204190 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2 is running failed: container process not found" containerID="9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:53:08 crc kubenswrapper[4967]: E0120 08:53:08.204598 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2 is running failed: container process not found" containerID="9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:53:08 crc kubenswrapper[4967]: E0120 08:53:08.204740 4967 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-58rr8" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="registry-server" Jan 20 08:53:08 crc kubenswrapper[4967]: I0120 08:53:08.333461 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:53:09 crc kubenswrapper[4967]: I0120 08:53:09.055858 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9fwf7" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="registry-server" containerID="cri-o://ed13ef5eeb4ee4987cde6255c3e782c34c9f1dca4777a6a83dcc49d06839d6cc" gracePeriod=2 Jan 20 08:53:09 crc kubenswrapper[4967]: I0120 08:53:09.083383 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-jbrxm" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="registry-server" probeResult="failure" output=< Jan 20 08:53:09 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 08:53:09 crc kubenswrapper[4967]: > Jan 20 08:53:10 crc kubenswrapper[4967]: I0120 08:53:10.876269 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:53:10 crc kubenswrapper[4967]: I0120 08:53:10.912326 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.026828 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.122635 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89thw\" (UniqueName: \"kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw\") pod \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.122806 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities\") pod \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.122866 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content\") pod \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\" (UID: \"1171240e-d46a-4eca-b9dc-fcc9e14b27a0\") " Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.124666 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities" (OuterVolumeSpecName: "utilities") pod "1171240e-d46a-4eca-b9dc-fcc9e14b27a0" (UID: "1171240e-d46a-4eca-b9dc-fcc9e14b27a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.129560 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw" (OuterVolumeSpecName: "kube-api-access-89thw") pod "1171240e-d46a-4eca-b9dc-fcc9e14b27a0" (UID: "1171240e-d46a-4eca-b9dc-fcc9e14b27a0"). InnerVolumeSpecName "kube-api-access-89thw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.173825 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1171240e-d46a-4eca-b9dc-fcc9e14b27a0" (UID: "1171240e-d46a-4eca-b9dc-fcc9e14b27a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.217506 4967 generic.go:334] "Generic (PLEG): container finished" podID="20939b87-fa8e-4128-be15-edbd83a634b6" containerID="ed13ef5eeb4ee4987cde6255c3e782c34c9f1dca4777a6a83dcc49d06839d6cc" exitCode=0 Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.217563 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerDied","Data":"ed13ef5eeb4ee4987cde6255c3e782c34c9f1dca4777a6a83dcc49d06839d6cc"} Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.220024 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58rr8" event={"ID":"1171240e-d46a-4eca-b9dc-fcc9e14b27a0","Type":"ContainerDied","Data":"ffe8e4b7782f1863c79b4962b00b43cd2b7ec7ccb472f940d3d29f93e0a2009a"} Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.220062 4967 scope.go:117] "RemoveContainer" containerID="9ae814cd2c3b9e40bb10f08ee66102c1aa8cb28a22c676316ba87e5ca751d9b2" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.220166 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58rr8" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.225335 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.225520 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.225662 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89thw\" (UniqueName: \"kubernetes.io/projected/1171240e-d46a-4eca-b9dc-fcc9e14b27a0-kube-api-access-89thw\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.248690 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.253207 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-58rr8"] Jan 20 08:53:13 crc kubenswrapper[4967]: I0120 08:53:13.706028 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" path="/var/lib/kubelet/pods/1171240e-d46a-4eca-b9dc-fcc9e14b27a0/volumes" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.018717 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.136493 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7x8h\" (UniqueName: \"kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h\") pod \"20939b87-fa8e-4128-be15-edbd83a634b6\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.136598 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities\") pod \"20939b87-fa8e-4128-be15-edbd83a634b6\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.136689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content\") pod \"20939b87-fa8e-4128-be15-edbd83a634b6\" (UID: \"20939b87-fa8e-4128-be15-edbd83a634b6\") " Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.138254 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities" (OuterVolumeSpecName: "utilities") pod "20939b87-fa8e-4128-be15-edbd83a634b6" (UID: "20939b87-fa8e-4128-be15-edbd83a634b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.142841 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h" (OuterVolumeSpecName: "kube-api-access-n7x8h") pod "20939b87-fa8e-4128-be15-edbd83a634b6" (UID: "20939b87-fa8e-4128-be15-edbd83a634b6"). InnerVolumeSpecName "kube-api-access-n7x8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.190971 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20939b87-fa8e-4128-be15-edbd83a634b6" (UID: "20939b87-fa8e-4128-be15-edbd83a634b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.227562 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fwf7" event={"ID":"20939b87-fa8e-4128-be15-edbd83a634b6","Type":"ContainerDied","Data":"75b0b98f6e44493351bc7d52b49657cd6a33c0c0c4a81ed8ef54b13bf323219d"} Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.227669 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fwf7" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.238568 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7x8h\" (UniqueName: \"kubernetes.io/projected/20939b87-fa8e-4128-be15-edbd83a634b6-kube-api-access-n7x8h\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.238617 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.238630 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20939b87-fa8e-4128-be15-edbd83a634b6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.251580 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:53:14 crc kubenswrapper[4967]: I0120 08:53:14.254135 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9fwf7"] Jan 20 08:53:15 crc kubenswrapper[4967]: I0120 08:53:15.708828 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" path="/var/lib/kubelet/pods/20939b87-fa8e-4128-be15-edbd83a634b6/volumes" Jan 20 08:53:15 crc kubenswrapper[4967]: I0120 08:53:15.896652 4967 scope.go:117] "RemoveContainer" containerID="ef2500f511c5ebecf19d7d823fd4988223d863a09186bb490c81f18f7be52c6c" Jan 20 08:53:16 crc kubenswrapper[4967]: I0120 08:53:16.175032 4967 scope.go:117] "RemoveContainer" containerID="f504f912811c35675f7fb315da1e265e39c046bc7b90bd2d743d00e099ba2c98" Jan 20 08:53:16 crc kubenswrapper[4967]: I0120 08:53:16.974826 4967 scope.go:117] "RemoveContainer" containerID="ed13ef5eeb4ee4987cde6255c3e782c34c9f1dca4777a6a83dcc49d06839d6cc" Jan 20 08:53:17 crc kubenswrapper[4967]: I0120 08:53:16.999899 4967 scope.go:117] "RemoveContainer" containerID="8dc8aa7c49e8862a71bef2e6cb613a4697210c9192d37dc9dcf4f516f0314e9b" Jan 20 08:53:17 crc kubenswrapper[4967]: I0120 08:53:17.034429 4967 scope.go:117] "RemoveContainer" containerID="106f0f70565bffb32765d09698883dd6c4e55ceb27e255f9fcc9ab01500bc6e6" Jan 20 08:53:17 crc kubenswrapper[4967]: I0120 08:53:17.681258 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.251053 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerStarted","Data":"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307"} Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.252783 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerStarted","Data":"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41"} Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.254181 4967 generic.go:334] "Generic (PLEG): container finished" podID="551f69d7-2c12-4474-9169-4540933734e1" containerID="ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1" exitCode=0 Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.254205 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerDied","Data":"ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1"} Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.275398 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9pxql" podStartSLOduration=5.252199451 podStartE2EDuration="1m19.275377484s" podCreationTimestamp="2026-01-20 08:51:59 +0000 UTC" firstStartedPulling="2026-01-20 08:52:01.241537519 +0000 UTC m=+155.826837726" lastFinishedPulling="2026-01-20 08:53:15.264715552 +0000 UTC m=+229.850015759" observedRunningTime="2026-01-20 08:53:18.274305909 +0000 UTC m=+232.859606126" watchObservedRunningTime="2026-01-20 08:53:18.275377484 +0000 UTC m=+232.860677691" Jan 20 08:53:18 crc kubenswrapper[4967]: I0120 08:53:18.290829 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pn7jw" podStartSLOduration=4.681233586 podStartE2EDuration="1m18.290811834s" podCreationTimestamp="2026-01-20 08:52:00 +0000 UTC" firstStartedPulling="2026-01-20 08:52:02.287115615 +0000 UTC m=+156.872415822" lastFinishedPulling="2026-01-20 08:53:15.896693863 +0000 UTC m=+230.481994070" observedRunningTime="2026-01-20 08:53:18.288528968 +0000 UTC m=+232.873829175" watchObservedRunningTime="2026-01-20 08:53:18.290811834 +0000 UTC m=+232.876112041" Jan 20 08:53:19 crc kubenswrapper[4967]: I0120 08:53:19.261851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerStarted","Data":"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9"} Jan 20 08:53:19 crc kubenswrapper[4967]: I0120 08:53:19.284881 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d9b7s" podStartSLOduration=2.570655609 podStartE2EDuration="1m20.284862365s" podCreationTimestamp="2026-01-20 08:51:59 +0000 UTC" firstStartedPulling="2026-01-20 08:52:01.259440291 +0000 UTC m=+155.844740498" lastFinishedPulling="2026-01-20 08:53:18.973647047 +0000 UTC m=+233.558947254" observedRunningTime="2026-01-20 08:53:19.282515857 +0000 UTC m=+233.867816064" watchObservedRunningTime="2026-01-20 08:53:19.284862365 +0000 UTC m=+233.870162592" Jan 20 08:53:19 crc kubenswrapper[4967]: I0120 08:53:19.645387 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-57z2c"] Jan 20 08:53:19 crc kubenswrapper[4967]: I0120 08:53:19.793508 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:53:19 crc kubenswrapper[4967]: I0120 08:53:19.793555 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:53:20 crc kubenswrapper[4967]: I0120 08:53:20.205649 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:53:20 crc kubenswrapper[4967]: I0120 08:53:20.205867 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:53:20 crc kubenswrapper[4967]: I0120 08:53:20.243165 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:53:20 crc kubenswrapper[4967]: I0120 08:53:20.977903 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-d9b7s" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" probeResult="failure" output=< Jan 20 08:53:20 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 08:53:20 crc kubenswrapper[4967]: > Jan 20 08:53:21 crc kubenswrapper[4967]: I0120 08:53:21.238168 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:53:21 crc kubenswrapper[4967]: I0120 08:53:21.238349 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.279694 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pn7jw" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="registry-server" probeResult="failure" output=< Jan 20 08:53:22 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 08:53:22 crc kubenswrapper[4967]: > Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.522969 4967 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523500 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b38ddc-1951-4356-8b18-f6d32a883fd0" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523526 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b38ddc-1951-4356-8b18-f6d32a883fd0" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523557 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523592 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523632 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="extract-content" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523641 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="extract-content" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523657 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="extract-content" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523666 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="extract-content" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523684 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523692 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523711 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c574c9-8911-4502-bd88-d3c530e26201" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523719 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c574c9-8911-4502-bd88-d3c530e26201" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523744 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="extract-utilities" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523752 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="extract-utilities" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.523772 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="extract-utilities" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.523781 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="extract-utilities" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.524041 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b38ddc-1951-4356-8b18-f6d32a883fd0" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.524061 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="31c574c9-8911-4502-bd88-d3c530e26201" containerName="pruner" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.524080 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="20939b87-fa8e-4128-be15-edbd83a634b6" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.524100 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1171240e-d46a-4eca-b9dc-fcc9e14b27a0" containerName="registry-server" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.525072 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.549371 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.549427 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.549471 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.549499 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.549751 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.557064 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.585914 4967 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.586281 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb" gracePeriod=15 Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.586325 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781" gracePeriod=15 Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.586388 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12" gracePeriod=15 Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.586397 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f" gracePeriod=15 Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.586440 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93" gracePeriod=15 Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588025 4967 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588311 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588387 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588451 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588507 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588565 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588638 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588703 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588762 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588822 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588873 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.588935 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.588985 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589140 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589219 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589275 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589328 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589386 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589446 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.589585 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.589680 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650710 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650768 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650811 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650842 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650883 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650919 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650969 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.650982 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.651015 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.651040 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.651070 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.651109 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.751828 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.752231 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.751972 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.752281 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.752383 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.752379 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: I0120 08:53:22.853314 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:53:22 crc kubenswrapper[4967]: W0120 08:53:22.878913 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-eb2b0d614c44fac4a0fd8c40c87dbb1537fe77ff1a3e30dc84a011e776a34ca2 WatchSource:0}: Error finding container eb2b0d614c44fac4a0fd8c40c87dbb1537fe77ff1a3e30dc84a011e776a34ca2: Status 404 returned error can't find the container with id eb2b0d614c44fac4a0fd8c40c87dbb1537fe77ff1a3e30dc84a011e776a34ca2 Jan 20 08:53:22 crc kubenswrapper[4967]: E0120 08:53:22.883113 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188c6473f7474e67 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-20 08:53:22.882424423 +0000 UTC m=+237.467724630,LastTimestamp:2026-01-20 08:53:22.882424423 +0000 UTC m=+237.467724630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.283143 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.284438 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.285261 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93" exitCode=0 Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.285295 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f" exitCode=0 Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.285309 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781" exitCode=0 Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.285321 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12" exitCode=2 Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.285364 4967 scope.go:117] "RemoveContainer" containerID="0721faed7bf8f419387b2c59ed56affb4eda9154190d593d301ded1d7ef32ab4" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.288315 4967 generic.go:334] "Generic (PLEG): container finished" podID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" containerID="a58c523040356958f654c1c6799a72ca5565e08e5fb5cffdba2d3c4e3e42dd5b" exitCode=0 Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.288357 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2199d68b-78b9-4479-b622-8c5d7fc2f1c8","Type":"ContainerDied","Data":"a58c523040356958f654c1c6799a72ca5565e08e5fb5cffdba2d3c4e3e42dd5b"} Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.289064 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.289368 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.289722 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1237d81834027cd4e18a1ca54195b028dcf405c79d987c6f5b78c650ff383cd9"} Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.289754 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"eb2b0d614c44fac4a0fd8c40c87dbb1537fe77ff1a3e30dc84a011e776a34ca2"} Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.289793 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.290182 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.290806 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:23 crc kubenswrapper[4967]: I0120 08:53:23.291082 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.298077 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.548405 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.549067 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.549496 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.574987 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access\") pod \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575049 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir\") pod \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575138 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock\") pod \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\" (UID: \"2199d68b-78b9-4479-b622-8c5d7fc2f1c8\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575221 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2199d68b-78b9-4479-b622-8c5d7fc2f1c8" (UID: "2199d68b-78b9-4479-b622-8c5d7fc2f1c8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575291 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock" (OuterVolumeSpecName: "var-lock") pod "2199d68b-78b9-4479-b622-8c5d7fc2f1c8" (UID: "2199d68b-78b9-4479-b622-8c5d7fc2f1c8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575652 4967 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-var-lock\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.575679 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.579091 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2199d68b-78b9-4479-b622-8c5d7fc2f1c8" (UID: "2199d68b-78b9-4479-b622-8c5d7fc2f1c8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.677286 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2199d68b-78b9-4479-b622-8c5d7fc2f1c8-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.952118 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.953100 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.953763 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.954493 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.955051 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.981992 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982094 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982091 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982163 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982175 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982201 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982502 4967 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982527 4967 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:24 crc kubenswrapper[4967]: I0120 08:53:24.982541 4967 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.307304 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.309014 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb" exitCode=0 Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.309079 4967 scope.go:117] "RemoveContainer" containerID="1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.309183 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.311517 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2199d68b-78b9-4479-b622-8c5d7fc2f1c8","Type":"ContainerDied","Data":"16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c"} Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.311656 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16e8a63dfce62476846d146cc6679674efc6a63dc304bc2819aceaf7dacb784c" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.311569 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.324266 4967 scope.go:117] "RemoveContainer" containerID="8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.331225 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.331692 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.332166 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.336137 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.336406 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.336712 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.337422 4967 scope.go:117] "RemoveContainer" containerID="452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.350374 4967 scope.go:117] "RemoveContainer" containerID="176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.363626 4967 scope.go:117] "RemoveContainer" containerID="574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.378095 4967 scope.go:117] "RemoveContainer" containerID="e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.399771 4967 scope.go:117] "RemoveContainer" containerID="1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.400272 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\": container with ID starting with 1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93 not found: ID does not exist" containerID="1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.400324 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93"} err="failed to get container status \"1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\": rpc error: code = NotFound desc = could not find container \"1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93\": container with ID starting with 1e506c2926ef7ce81dd1681c18564d2862e588c27b4951e0b71b676f8430bc93 not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.400357 4967 scope.go:117] "RemoveContainer" containerID="8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.400804 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\": container with ID starting with 8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f not found: ID does not exist" containerID="8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.400831 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f"} err="failed to get container status \"8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\": rpc error: code = NotFound desc = could not find container \"8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f\": container with ID starting with 8e9201d6d720c7ea0cc5fd4334b0d83ca7a0e62191b92f963b228e33d884b85f not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.400853 4967 scope.go:117] "RemoveContainer" containerID="452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.401214 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\": container with ID starting with 452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781 not found: ID does not exist" containerID="452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.401240 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781"} err="failed to get container status \"452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\": rpc error: code = NotFound desc = could not find container \"452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781\": container with ID starting with 452d3a9c69b71232ca4f690d226049def4d4b98af082f6d3e1296bcc3186e781 not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.401259 4967 scope.go:117] "RemoveContainer" containerID="176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.401564 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\": container with ID starting with 176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12 not found: ID does not exist" containerID="176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.401598 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12"} err="failed to get container status \"176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\": rpc error: code = NotFound desc = could not find container \"176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12\": container with ID starting with 176d2eab28516c2303041f5995e10bfc1179bfa00b0afc71f1300f11f5e98d12 not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.401639 4967 scope.go:117] "RemoveContainer" containerID="574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.402522 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\": container with ID starting with 574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb not found: ID does not exist" containerID="574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.402601 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb"} err="failed to get container status \"574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\": rpc error: code = NotFound desc = could not find container \"574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb\": container with ID starting with 574dad29603fc4a21938e00b7af0a8d57207fd7223640f45714550882c20aabb not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.402689 4967 scope.go:117] "RemoveContainer" containerID="e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892" Jan 20 08:53:25 crc kubenswrapper[4967]: E0120 08:53:25.403298 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\": container with ID starting with e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892 not found: ID does not exist" containerID="e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.403342 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892"} err="failed to get container status \"e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\": rpc error: code = NotFound desc = could not find container \"e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892\": container with ID starting with e0c6bd3aadbd18b49871986288a490107f3e1849a7290a6b0428a556ea09c892 not found: ID does not exist" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.704446 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.704844 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.705175 4967 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:25 crc kubenswrapper[4967]: I0120 08:53:25.711516 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 20 08:53:29 crc kubenswrapper[4967]: E0120 08:53:29.459779 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188c6473f7474e67 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-20 08:53:22.882424423 +0000 UTC m=+237.467724630,LastTimestamp:2026-01-20 08:53:22.882424423 +0000 UTC m=+237.467724630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.842073 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.843115 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.843592 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.844049 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.887910 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.888414 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.888886 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:29 crc kubenswrapper[4967]: I0120 08:53:29.889317 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.267092 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.267935 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.268799 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.269331 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.269717 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.743561 4967 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" volumeName="registry-storage" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.949689 4967 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.950367 4967 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.950879 4967 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.951372 4967 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.951890 4967 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:30 crc kubenswrapper[4967]: I0120 08:53:30.951942 4967 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 20 08:53:30 crc kubenswrapper[4967]: E0120 08:53:30.952325 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="200ms" Jan 20 08:53:31 crc kubenswrapper[4967]: E0120 08:53:31.153746 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="400ms" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.278861 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.280367 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.281192 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.281597 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.282347 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.282978 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.326768 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.327360 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.327763 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.328181 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.328862 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: I0120 08:53:31.329222 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:31 crc kubenswrapper[4967]: E0120 08:53:31.555001 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="800ms" Jan 20 08:53:32 crc kubenswrapper[4967]: E0120 08:53:32.356135 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="1.6s" Jan 20 08:53:33 crc kubenswrapper[4967]: E0120 08:53:33.957088 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="3.2s" Jan 20 08:53:35 crc kubenswrapper[4967]: I0120 08:53:35.697405 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:35 crc kubenswrapper[4967]: I0120 08:53:35.697894 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:35 crc kubenswrapper[4967]: I0120 08:53:35.698274 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:35 crc kubenswrapper[4967]: I0120 08:53:35.698831 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:35 crc kubenswrapper[4967]: I0120 08:53:35.699115 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.375735 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.376150 4967 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7" exitCode=1 Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.376179 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7"} Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.376558 4967 scope.go:117] "RemoveContainer" containerID="13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.377102 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.377495 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.377757 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.378000 4967 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.378257 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:36 crc kubenswrapper[4967]: I0120 08:53:36.378449 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: E0120 08:53:37.157742 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="6.4s" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.386408 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.386483 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c80b331d130a8fc49347891466e82b7c21724b9e1207f9dbed661766b5565cdf"} Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.387664 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.388212 4967 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.388715 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.389179 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.389509 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.389964 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.693312 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.694206 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.694778 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.695121 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.695472 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.695847 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.696257 4967 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.706950 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.706982 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:37 crc kubenswrapper[4967]: E0120 08:53:37.707309 4967 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:37 crc kubenswrapper[4967]: I0120 08:53:37.707802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:37 crc kubenswrapper[4967]: W0120 08:53:37.726337 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-d2da26838642360a6e452641fcc3254d9a565fd18a0015c5ed0faaec10a8187b WatchSource:0}: Error finding container d2da26838642360a6e452641fcc3254d9a565fd18a0015c5ed0faaec10a8187b: Status 404 returned error can't find the container with id d2da26838642360a6e452641fcc3254d9a565fd18a0015c5ed0faaec10a8187b Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.393060 4967 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d0114bdd2f8bc64d562e58810dabf76a805196cd33f3eca8de8fe75b2186ef79" exitCode=0 Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.393360 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d0114bdd2f8bc64d562e58810dabf76a805196cd33f3eca8de8fe75b2186ef79"} Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.393389 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d2da26838642360a6e452641fcc3254d9a565fd18a0015c5ed0faaec10a8187b"} Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.393639 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.393654 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:38 crc kubenswrapper[4967]: E0120 08:53:38.394232 4967 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.394467 4967 status_manager.go:851] "Failed to get status for pod" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.394976 4967 status_manager.go:851] "Failed to get status for pod" podUID="551f69d7-2c12-4474-9169-4540933734e1" pod="openshift-marketplace/redhat-marketplace-d9b7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-d9b7s\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.395360 4967 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.395836 4967 status_manager.go:851] "Failed to get status for pod" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" pod="openshift-marketplace/redhat-operators-pn7jw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pn7jw\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.396121 4967 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.396417 4967 status_manager.go:851] "Failed to get status for pod" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" pod="openshift-marketplace/redhat-marketplace-9pxql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-9pxql\": dial tcp 38.102.83.136:6443: connect: connection refused" Jan 20 08:53:38 crc kubenswrapper[4967]: I0120 08:53:38.662068 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:53:39 crc kubenswrapper[4967]: I0120 08:53:39.420129 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd4185f1494315e131124d04d1bea1f249fade85d27954cdc57e7b15a89dbb21"} Jan 20 08:53:39 crc kubenswrapper[4967]: I0120 08:53:39.420427 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"881cd36e119f349d5895df5f2412962e70eed988a7fddfd838a7ead351534252"} Jan 20 08:53:39 crc kubenswrapper[4967]: I0120 08:53:39.420438 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"93d5b9dc2b25f50f59e0b04b1bc8f244dd2ec91ea2cf155b8ee1126bea57d43e"} Jan 20 08:53:39 crc kubenswrapper[4967]: I0120 08:53:39.420446 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4669167a6c13e9b727e92476f7ed67c584ed0c0d841292d15416b154aa760155"} Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.447808 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef57aee159655227fb7408e9e58a024c63c33e3246f30901d6b5a761a783e1ad"} Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.448205 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.451317 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.451354 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.459800 4967 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.840257 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.840365 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 20 08:53:41 crc kubenswrapper[4967]: I0120 08:53:41.840424 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 20 08:53:42 crc kubenswrapper[4967]: I0120 08:53:42.453268 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:42 crc kubenswrapper[4967]: I0120 08:53:42.453306 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:42 crc kubenswrapper[4967]: I0120 08:53:42.708893 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:42 crc kubenswrapper[4967]: I0120 08:53:42.708938 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:42 crc kubenswrapper[4967]: I0120 08:53:42.713736 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:43 crc kubenswrapper[4967]: I0120 08:53:43.459513 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:43 crc kubenswrapper[4967]: I0120 08:53:43.459557 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:43 crc kubenswrapper[4967]: I0120 08:53:43.465474 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:44 crc kubenswrapper[4967]: I0120 08:53:44.464677 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:44 crc kubenswrapper[4967]: I0120 08:53:44.464998 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:44 crc kubenswrapper[4967]: I0120 08:53:44.665503 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" podUID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" containerName="oauth-openshift" containerID="cri-o://9f8171ab6a0fe188c4f98a89800bf41a0a4d5b2edd9f048a0b7fa93cb63cd050" gracePeriod=15 Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.474217 4967 generic.go:334] "Generic (PLEG): container finished" podID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" containerID="9f8171ab6a0fe188c4f98a89800bf41a0a4d5b2edd9f048a0b7fa93cb63cd050" exitCode=0 Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.474312 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" event={"ID":"7db3ffb6-ef1c-432b-8235-5ef6d77525a4","Type":"ContainerDied","Data":"9f8171ab6a0fe188c4f98a89800bf41a0a4d5b2edd9f048a0b7fa93cb63cd050"} Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.474669 4967 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.474684 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e03e4415-da9f-4306-8194-30d9bbc66a86" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.606513 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.705983 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706070 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706112 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706143 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706178 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706225 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkpjt\" (UniqueName: \"kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706226 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706269 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706355 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706396 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706460 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706520 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706584 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706652 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig\") pod \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\" (UID: \"7db3ffb6-ef1c-432b-8235-5ef6d77525a4\") " Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.706964 4967 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.707299 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.707317 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.708095 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.709204 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.713358 4967 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d9847098-7851-4b9e-addc-1d09a8c3c94f" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.713757 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.714070 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.715440 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.715790 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt" (OuterVolumeSpecName: "kube-api-access-lkpjt") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "kube-api-access-lkpjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.716356 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.716598 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.716935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.717053 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.719396 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7db3ffb6-ef1c-432b-8235-5ef6d77525a4" (UID: "7db3ffb6-ef1c-432b-8235-5ef6d77525a4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.810412 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811052 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811076 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811091 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811103 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811114 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811126 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811139 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811149 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811159 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811172 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811184 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:45 crc kubenswrapper[4967]: I0120 08:53:45.811197 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkpjt\" (UniqueName: \"kubernetes.io/projected/7db3ffb6-ef1c-432b-8235-5ef6d77525a4-kube-api-access-lkpjt\") on node \"crc\" DevicePath \"\"" Jan 20 08:53:46 crc kubenswrapper[4967]: I0120 08:53:46.489512 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" event={"ID":"7db3ffb6-ef1c-432b-8235-5ef6d77525a4","Type":"ContainerDied","Data":"98d92dbdbdbe21f3f06c41e2055aae90eb7ec4fb07eb5731e51c371ce2992529"} Jan 20 08:53:46 crc kubenswrapper[4967]: I0120 08:53:46.489560 4967 scope.go:117] "RemoveContainer" containerID="9f8171ab6a0fe188c4f98a89800bf41a0a4d5b2edd9f048a0b7fa93cb63cd050" Jan 20 08:53:46 crc kubenswrapper[4967]: I0120 08:53:46.489679 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-57z2c" Jan 20 08:53:51 crc kubenswrapper[4967]: I0120 08:53:51.843561 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 20 08:53:51 crc kubenswrapper[4967]: I0120 08:53:51.844266 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 20 08:53:54 crc kubenswrapper[4967]: I0120 08:53:54.946726 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 20 08:53:54 crc kubenswrapper[4967]: I0120 08:53:54.976378 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.037603 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.113449 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.215957 4967 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.220989 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=33.220967381 podStartE2EDuration="33.220967381s" podCreationTimestamp="2026-01-20 08:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:53:41.560146268 +0000 UTC m=+256.145446475" watchObservedRunningTime="2026-01-20 08:53:55.220967381 +0000 UTC m=+269.806267608" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.221810 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-57z2c","openshift-kube-apiserver/kube-apiserver-crc"] Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.221885 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.229461 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.250742 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.250716275 podStartE2EDuration="14.250716275s" podCreationTimestamp="2026-01-20 08:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:53:55.247405204 +0000 UTC m=+269.832705451" watchObservedRunningTime="2026-01-20 08:53:55.250716275 +0000 UTC m=+269.836016502" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.329832 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.678734 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.711459 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" path="/var/lib/kubelet/pods/7db3ffb6-ef1c-432b-8235-5ef6d77525a4/volumes" Jan 20 08:53:55 crc kubenswrapper[4967]: I0120 08:53:55.810554 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.177546 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.220717 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.232386 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.233291 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.653506 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.730541 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.875099 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 20 08:53:56 crc kubenswrapper[4967]: I0120 08:53:56.995215 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.116833 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.228950 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.263835 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.628146 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.769520 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 20 08:53:57 crc kubenswrapper[4967]: I0120 08:53:57.780305 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.004857 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.036291 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.045993 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.150701 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.197438 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.198379 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.229904 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.309099 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.388417 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.439052 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.450596 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.496297 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.540514 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.724183 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.735397 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.747354 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.805840 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.807308 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.842765 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.842970 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.924661 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 20 08:53:58 crc kubenswrapper[4967]: I0120 08:53:58.990955 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.055460 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.065635 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.066514 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.091510 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.094508 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.105607 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.185237 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.190444 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.239396 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.304803 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.453988 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.572052 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.577980 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.634351 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.680334 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.701115 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.809406 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.882132 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.943351 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 20 08:53:59 crc kubenswrapper[4967]: I0120 08:53:59.951378 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.007897 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.024232 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.049918 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.216373 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.232970 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.238007 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.314232 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.418273 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.450030 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.466378 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.508854 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.537400 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.729518 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.782172 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.806802 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.849666 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 20 08:54:00 crc kubenswrapper[4967]: I0120 08:54:00.857939 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.051124 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.130591 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.195026 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.195842 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.231593 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.367024 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.367347 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.401796 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.442811 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.454671 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.607002 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.611429 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.629219 4967 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.647461 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.841174 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.841264 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.841391 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.842314 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"c80b331d130a8fc49347891466e82b7c21724b9e1207f9dbed661766b5565cdf"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.842520 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://c80b331d130a8fc49347891466e82b7c21724b9e1207f9dbed661766b5565cdf" gracePeriod=30 Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.860808 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.932480 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7544d6d989-kgr9n"] Jan 20 08:54:01 crc kubenswrapper[4967]: E0120 08:54:01.932816 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" containerName="oauth-openshift" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.932842 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" containerName="oauth-openshift" Jan 20 08:54:01 crc kubenswrapper[4967]: E0120 08:54:01.932875 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" containerName="installer" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.932888 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" containerName="installer" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.933540 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2199d68b-78b9-4479-b622-8c5d7fc2f1c8" containerName="installer" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.933582 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db3ffb6-ef1c-432b-8235-5ef6d77525a4" containerName="oauth-openshift" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.934648 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.936944 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.937081 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.937282 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.937460 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.938124 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.938516 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941012 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941038 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941056 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgzs\" (UniqueName: \"kubernetes.io/projected/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-kube-api-access-wdgzs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941083 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-dir\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941106 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-session\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941130 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-policies\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941152 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941171 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941199 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-service-ca\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941222 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-login\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941256 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941285 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-router-certs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941326 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941360 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941372 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.941395 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-error\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.943167 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7544d6d989-kgr9n"] Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.943891 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.944082 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.949424 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.950043 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.959262 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.966997 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.973014 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.975299 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 20 08:54:01 crc kubenswrapper[4967]: I0120 08:54:01.988390 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.028915 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042453 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgzs\" (UniqueName: \"kubernetes.io/projected/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-kube-api-access-wdgzs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042505 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-dir\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042536 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-session\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042565 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-policies\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042595 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042640 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042666 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-service-ca\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042690 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-login\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042724 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-router-certs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042767 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042777 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-dir\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042793 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.042982 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-error\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.043122 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.043756 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-service-ca\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.044302 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.044759 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-audit-policies\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.045938 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.049491 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-error\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.049593 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.050047 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-login\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.050434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.050740 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-session\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.051069 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-router-certs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.052916 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.053914 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.065835 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgzs\" (UniqueName: \"kubernetes.io/projected/2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800-kube-api-access-wdgzs\") pod \"oauth-openshift-7544d6d989-kgr9n\" (UID: \"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800\") " pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.075390 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.138231 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.274400 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.354911 4967 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.366718 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.389974 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.548937 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.578599 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.588273 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.645685 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.659314 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.713768 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7544d6d989-kgr9n"] Jan 20 08:54:02 crc kubenswrapper[4967]: W0120 08:54:02.720424 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fe2bdfc_9c4d_4ed7_a958_bec6cf5a2800.slice/crio-1b7c254e3472ba407f5c1e54d0f88a98d2a5b7525137e7a590ceb09db7e4a78b WatchSource:0}: Error finding container 1b7c254e3472ba407f5c1e54d0f88a98d2a5b7525137e7a590ceb09db7e4a78b: Status 404 returned error can't find the container with id 1b7c254e3472ba407f5c1e54d0f88a98d2a5b7525137e7a590ceb09db7e4a78b Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.813336 4967 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.839263 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.846252 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.849998 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.879013 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.879819 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 20 08:54:02 crc kubenswrapper[4967]: I0120 08:54:02.958782 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.051454 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.191997 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.202099 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.203422 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.217169 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.268798 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.287235 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.330190 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.372248 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.424544 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.431330 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.462922 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.464519 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.488875 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.525915 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.600990 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" event={"ID":"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800","Type":"ContainerStarted","Data":"5537bf8ad3f1eec5761efa9ba891acfe32c437950f6513c8460bce287446322e"} Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.601038 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" event={"ID":"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800","Type":"ContainerStarted","Data":"1b7c254e3472ba407f5c1e54d0f88a98d2a5b7525137e7a590ceb09db7e4a78b"} Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.601220 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.621558 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" podStartSLOduration=44.621540683 podStartE2EDuration="44.621540683s" podCreationTimestamp="2026-01-20 08:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:54:03.619457536 +0000 UTC m=+278.204757743" watchObservedRunningTime="2026-01-20 08:54:03.621540683 +0000 UTC m=+278.206840890" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.646542 4967 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.732492 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.816049 4967 patch_prober.go:28] interesting pod/oauth-openshift-7544d6d989-kgr9n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:55700->10.217.0.56:6443: read: connection reset by peer" start-of-body= Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.816114 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" podUID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:55700->10.217.0.56:6443: read: connection reset by peer" Jan 20 08:54:03 crc kubenswrapper[4967]: I0120 08:54:03.941156 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.201752 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.251667 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.297564 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.387324 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.395912 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.459774 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.539653 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.558953 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.607943 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7544d6d989-kgr9n_2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800/oauth-openshift/0.log" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.608172 4967 generic.go:334] "Generic (PLEG): container finished" podID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" containerID="5537bf8ad3f1eec5761efa9ba891acfe32c437950f6513c8460bce287446322e" exitCode=255 Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.608210 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" event={"ID":"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800","Type":"ContainerDied","Data":"5537bf8ad3f1eec5761efa9ba891acfe32c437950f6513c8460bce287446322e"} Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.609087 4967 scope.go:117] "RemoveContainer" containerID="5537bf8ad3f1eec5761efa9ba891acfe32c437950f6513c8460bce287446322e" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.647028 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.821147 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.853090 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.857596 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.903223 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 20 08:54:04 crc kubenswrapper[4967]: I0120 08:54:04.982318 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.031061 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.069552 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.127562 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.242561 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.391137 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.488292 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 20 08:54:05 crc kubenswrapper[4967]: I0120 08:54:05.557395 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.618556 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7544d6d989-kgr9n_2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800/oauth-openshift/1.log" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.619337 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7544d6d989-kgr9n_2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800/oauth-openshift/0.log" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.619379 4967 generic.go:334] "Generic (PLEG): container finished" podID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" containerID="03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723" exitCode=255 Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.619405 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" event={"ID":"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800","Type":"ContainerDied","Data":"03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723"} Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.619435 4967 scope.go:117] "RemoveContainer" containerID="5537bf8ad3f1eec5761efa9ba891acfe32c437950f6513c8460bce287446322e" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.620229 4967 scope.go:117] "RemoveContainer" containerID="03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723" Jan 20 08:54:06 crc kubenswrapper[4967]: E0120 08:54:05.620823 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7544d6d989-kgr9n_openshift-authentication(2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800)\"" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" podUID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.632461 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.701305 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.703284 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.771088 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.773254 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.774866 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.789899 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.826567 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.863071 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.864018 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:05.971300 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.115894 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.182737 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.223676 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.232521 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.349222 4967 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.349424 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://1237d81834027cd4e18a1ca54195b028dcf405c79d987c6f5b78c650ff383cd9" gracePeriod=5 Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.355185 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.391891 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.429770 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.455047 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.471293 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.527177 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.591841 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.625989 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7544d6d989-kgr9n_2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800/oauth-openshift/1.log" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.626788 4967 scope.go:117] "RemoveContainer" containerID="03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723" Jan 20 08:54:06 crc kubenswrapper[4967]: E0120 08:54:06.627172 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7544d6d989-kgr9n_openshift-authentication(2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800)\"" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" podUID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.714750 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.791937 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.806510 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.828381 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.860980 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 20 08:54:06 crc kubenswrapper[4967]: I0120 08:54:06.919510 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.029751 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.038897 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.048108 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.089749 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.119440 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.124579 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.160057 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.168219 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.179789 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.224284 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.699029 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.758356 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.925668 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.925687 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.965046 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 20 08:54:07 crc kubenswrapper[4967]: I0120 08:54:07.996288 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.109472 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.211778 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.354221 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.375419 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.451010 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.548000 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.604696 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.631301 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.769792 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.865239 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.879264 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 20 08:54:08 crc kubenswrapper[4967]: I0120 08:54:08.976396 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.061474 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.246817 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.358379 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.366369 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.560348 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.601528 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.927383 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.941174 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 20 08:54:09 crc kubenswrapper[4967]: I0120 08:54:09.947298 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.072902 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.211134 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.280434 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.433839 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.464792 4967 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.591812 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.592454 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.715448 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 20 08:54:10 crc kubenswrapper[4967]: I0120 08:54:10.942720 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.303734 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.653741 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.653795 4967 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="1237d81834027cd4e18a1ca54195b028dcf405c79d987c6f5b78c650ff383cd9" exitCode=137 Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.938069 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.986818 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.986896 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990233 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990278 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990310 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990375 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990411 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990489 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990541 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990527 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.990632 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.991029 4967 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.991056 4967 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.991074 4967 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.991089 4967 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 20 08:54:11 crc kubenswrapper[4967]: I0120 08:54:11.999930 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.092815 4967 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.275019 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.275284 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.276079 4967 scope.go:117] "RemoveContainer" containerID="03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723" Jan 20 08:54:12 crc kubenswrapper[4967]: E0120 08:54:12.276487 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7544d6d989-kgr9n_openshift-authentication(2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800)\"" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" podUID="2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.631153 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.663480 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.663949 4967 scope.go:117] "RemoveContainer" containerID="1237d81834027cd4e18a1ca54195b028dcf405c79d987c6f5b78c650ff383cd9" Jan 20 08:54:12 crc kubenswrapper[4967]: I0120 08:54:12.663998 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.699902 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.700163 4967 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.710792 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.710830 4967 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="52bf0bb8-3276-4d19-9ce9-18d5f69e1222" Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.714451 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 20 08:54:13 crc kubenswrapper[4967]: I0120 08:54:13.714480 4967 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="52bf0bb8-3276-4d19-9ce9-18d5f69e1222" Jan 20 08:54:25 crc kubenswrapper[4967]: I0120 08:54:25.550471 4967 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 20 08:54:26 crc kubenswrapper[4967]: I0120 08:54:26.694885 4967 scope.go:117] "RemoveContainer" containerID="03cc330fd59233cc9b690ddcf9d0df64ffcaffa110f4e9a9886f47bc50304723" Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.756935 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7544d6d989-kgr9n_2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800/oauth-openshift/1.log" Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.757376 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" event={"ID":"2fe2bdfc-9c4d-4ed7-a958-bec6cf5a2800","Type":"ContainerStarted","Data":"8c7a121632429a8b6c8efe82f85ad140bcdd83b13b639a7fec381f47d901ad6d"} Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.758291 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.761755 4967 generic.go:334] "Generic (PLEG): container finished" podID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerID="fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11" exitCode=0 Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.761829 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerDied","Data":"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11"} Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.762946 4967 scope.go:117] "RemoveContainer" containerID="fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11" Jan 20 08:54:27 crc kubenswrapper[4967]: I0120 08:54:27.770847 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7544d6d989-kgr9n" Jan 20 08:54:28 crc kubenswrapper[4967]: I0120 08:54:28.775929 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerStarted","Data":"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275"} Jan 20 08:54:28 crc kubenswrapper[4967]: I0120 08:54:28.777057 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:54:28 crc kubenswrapper[4967]: I0120 08:54:28.779776 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.806091 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.807799 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.807833 4967 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c80b331d130a8fc49347891466e82b7c21724b9e1207f9dbed661766b5565cdf" exitCode=137 Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.807861 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c80b331d130a8fc49347891466e82b7c21724b9e1207f9dbed661766b5565cdf"} Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.807892 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"79292e773d943ed6c46b335e297b13dfdb15f6615ab218b6089d5132d64e7d0a"} Jan 20 08:54:32 crc kubenswrapper[4967]: I0120 08:54:32.807907 4967 scope.go:117] "RemoveContainer" containerID="13e418d38a04c51134a58827d70ce41664c5744a1932623b11a8cb5c7b8fe6a7" Jan 20 08:54:33 crc kubenswrapper[4967]: I0120 08:54:33.828209 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 20 08:54:34 crc kubenswrapper[4967]: I0120 08:54:34.276064 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 20 08:54:34 crc kubenswrapper[4967]: I0120 08:54:34.287693 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 20 08:54:34 crc kubenswrapper[4967]: I0120 08:54:34.643351 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 20 08:54:38 crc kubenswrapper[4967]: I0120 08:54:38.661573 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:54:41 crc kubenswrapper[4967]: I0120 08:54:41.841404 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:54:41 crc kubenswrapper[4967]: I0120 08:54:41.847748 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:54:46 crc kubenswrapper[4967]: I0120 08:54:46.886785 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 20 08:54:48 crc kubenswrapper[4967]: I0120 08:54:48.667625 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 20 08:54:49 crc kubenswrapper[4967]: I0120 08:54:49.979310 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.023957 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.024752 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" podUID="293c612b-323d-4408-af28-b9f559766583" containerName="controller-manager" containerID="cri-o://ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da" gracePeriod=30 Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.109792 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.110332 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" podUID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" containerName="route-controller-manager" containerID="cri-o://5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb" gracePeriod=30 Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.498286 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.634293 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtn6d\" (UniqueName: \"kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d\") pod \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.634348 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert\") pod \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.634414 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca\") pod \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.634459 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config\") pod \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\" (UID: \"ac422dc4-1356-4cbe-b1ec-d99946de9a13\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.635334 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca" (OuterVolumeSpecName: "client-ca") pod "ac422dc4-1356-4cbe-b1ec-d99946de9a13" (UID: "ac422dc4-1356-4cbe-b1ec-d99946de9a13"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.635496 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config" (OuterVolumeSpecName: "config") pod "ac422dc4-1356-4cbe-b1ec-d99946de9a13" (UID: "ac422dc4-1356-4cbe-b1ec-d99946de9a13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.635786 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.635804 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac422dc4-1356-4cbe-b1ec-d99946de9a13-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.642022 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ac422dc4-1356-4cbe-b1ec-d99946de9a13" (UID: "ac422dc4-1356-4cbe-b1ec-d99946de9a13"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.654552 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d" (OuterVolumeSpecName: "kube-api-access-xtn6d") pod "ac422dc4-1356-4cbe-b1ec-d99946de9a13" (UID: "ac422dc4-1356-4cbe-b1ec-d99946de9a13"). InnerVolumeSpecName "kube-api-access-xtn6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.737518 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtn6d\" (UniqueName: \"kubernetes.io/projected/ac422dc4-1356-4cbe-b1ec-d99946de9a13-kube-api-access-xtn6d\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.737577 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac422dc4-1356-4cbe-b1ec-d99946de9a13-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.826935 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.940190 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config\") pod \"293c612b-323d-4408-af28-b9f559766583\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.940274 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert\") pod \"293c612b-323d-4408-af28-b9f559766583\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.941045 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles\") pod \"293c612b-323d-4408-af28-b9f559766583\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.941383 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config" (OuterVolumeSpecName: "config") pod "293c612b-323d-4408-af28-b9f559766583" (UID: "293c612b-323d-4408-af28-b9f559766583"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.941725 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "293c612b-323d-4408-af28-b9f559766583" (UID: "293c612b-323d-4408-af28-b9f559766583"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.941888 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkbgz\" (UniqueName: \"kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz\") pod \"293c612b-323d-4408-af28-b9f559766583\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.941985 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca\") pod \"293c612b-323d-4408-af28-b9f559766583\" (UID: \"293c612b-323d-4408-af28-b9f559766583\") " Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.942484 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca" (OuterVolumeSpecName: "client-ca") pod "293c612b-323d-4408-af28-b9f559766583" (UID: "293c612b-323d-4408-af28-b9f559766583"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.942828 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.942857 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.942873 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/293c612b-323d-4408-af28-b9f559766583-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.944076 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "293c612b-323d-4408-af28-b9f559766583" (UID: "293c612b-323d-4408-af28-b9f559766583"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:55:05 crc kubenswrapper[4967]: I0120 08:55:05.944977 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz" (OuterVolumeSpecName: "kube-api-access-zkbgz") pod "293c612b-323d-4408-af28-b9f559766583" (UID: "293c612b-323d-4408-af28-b9f559766583"). InnerVolumeSpecName "kube-api-access-zkbgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.036735 4967 generic.go:334] "Generic (PLEG): container finished" podID="293c612b-323d-4408-af28-b9f559766583" containerID="ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da" exitCode=0 Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.036813 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" event={"ID":"293c612b-323d-4408-af28-b9f559766583","Type":"ContainerDied","Data":"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da"} Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.036845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" event={"ID":"293c612b-323d-4408-af28-b9f559766583","Type":"ContainerDied","Data":"a5088aed30a9aa770894dffdfef93a31bd8134cbc129044bee6572b5537799ba"} Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.036872 4967 scope.go:117] "RemoveContainer" containerID="ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.037852 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fbdgc" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.038509 4967 generic.go:334] "Generic (PLEG): container finished" podID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" containerID="5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb" exitCode=0 Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.038534 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" event={"ID":"ac422dc4-1356-4cbe-b1ec-d99946de9a13","Type":"ContainerDied","Data":"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb"} Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.038552 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" event={"ID":"ac422dc4-1356-4cbe-b1ec-d99946de9a13","Type":"ContainerDied","Data":"5765f60efb375c5a802963da33f7a96c172e76ce3745e0a28cf31287f953ed0c"} Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.038553 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.043873 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkbgz\" (UniqueName: \"kubernetes.io/projected/293c612b-323d-4408-af28-b9f559766583-kube-api-access-zkbgz\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.043908 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293c612b-323d-4408-af28-b9f559766583-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.052747 4967 scope.go:117] "RemoveContainer" containerID="ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da" Jan 20 08:55:06 crc kubenswrapper[4967]: E0120 08:55:06.053195 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da\": container with ID starting with ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da not found: ID does not exist" containerID="ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.053238 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da"} err="failed to get container status \"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da\": rpc error: code = NotFound desc = could not find container \"ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da\": container with ID starting with ba3dee2e63fbc80016586311b1dd4b7ddbe67ddd7f38d7da327a298e051a14da not found: ID does not exist" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.053301 4967 scope.go:117] "RemoveContainer" containerID="5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.059980 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.064633 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n544b"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.072982 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.073067 4967 scope.go:117] "RemoveContainer" containerID="5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb" Jan 20 08:55:06 crc kubenswrapper[4967]: E0120 08:55:06.073441 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb\": container with ID starting with 5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb not found: ID does not exist" containerID="5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.073471 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb"} err="failed to get container status \"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb\": rpc error: code = NotFound desc = could not find container \"5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb\": container with ID starting with 5590a8042af8bcb9ed9f75bb586f798137cab3328f67700d0cad062378ad8bcb not found: ID does not exist" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.075593 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fbdgc"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.234808 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:06 crc kubenswrapper[4967]: E0120 08:55:06.235399 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293c612b-323d-4408-af28-b9f559766583" containerName="controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.235498 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="293c612b-323d-4408-af28-b9f559766583" containerName="controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: E0120 08:55:06.235592 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" containerName="route-controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.235739 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" containerName="route-controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: E0120 08:55:06.235846 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.235929 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.236135 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.236221 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="293c612b-323d-4408-af28-b9f559766583" containerName="controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.236281 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" containerName="route-controller-manager" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.236779 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.239053 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.239469 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.239844 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.240095 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.240359 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.240727 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.248117 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.248212 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.248254 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.248309 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.248336 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75gr\" (UniqueName: \"kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.249495 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.259178 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.267888 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.268516 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.271902 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.272246 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.272578 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.275204 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.275391 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.275452 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.284909 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.350287 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.350339 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.350392 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.350415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75gr\" (UniqueName: \"kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.350446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.351426 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.351430 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.352350 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.354286 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.368704 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75gr\" (UniqueName: \"kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr\") pod \"controller-manager-5ff6ccf94b-nmhfl\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.451663 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.451714 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.451758 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6htws\" (UniqueName: \"kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.451783 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.554381 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.554509 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.554539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.554583 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6htws\" (UniqueName: \"kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.555930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.557587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.559569 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.566130 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.581959 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6htws\" (UniqueName: \"kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws\") pod \"route-controller-manager-dfff9545c-nfltr\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.585839 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.788304 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:06 crc kubenswrapper[4967]: I0120 08:55:06.950378 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:06 crc kubenswrapper[4967]: W0120 08:55:06.955620 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda429b03a_8a1d_479f_acd4_3aeebcc92d26.slice/crio-659148c3303728e54094f53b2cea6c735c2a3e6e2a8f57b6a8442f3b11731f8a WatchSource:0}: Error finding container 659148c3303728e54094f53b2cea6c735c2a3e6e2a8f57b6a8442f3b11731f8a: Status 404 returned error can't find the container with id 659148c3303728e54094f53b2cea6c735c2a3e6e2a8f57b6a8442f3b11731f8a Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.051242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" event={"ID":"e08d5d40-f3bd-446b-9687-1dc8420b92dd","Type":"ContainerStarted","Data":"758b56e72e521fb7d3f82d064d1e52ff11e3ef394dbcb1e05cd7d3ecb7bcf3a0"} Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.051298 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" event={"ID":"e08d5d40-f3bd-446b-9687-1dc8420b92dd","Type":"ContainerStarted","Data":"2e71f1496de1eaf6214dff57a8c1620fe44068592fb28845d2ec13e97db83781"} Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.052406 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.053249 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" event={"ID":"a429b03a-8a1d-479f-acd4-3aeebcc92d26","Type":"ContainerStarted","Data":"659148c3303728e54094f53b2cea6c735c2a3e6e2a8f57b6a8442f3b11731f8a"} Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.070466 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" podStartSLOduration=1.070446688 podStartE2EDuration="1.070446688s" podCreationTimestamp="2026-01-20 08:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:07.065436412 +0000 UTC m=+341.650736639" watchObservedRunningTime="2026-01-20 08:55:07.070446688 +0000 UTC m=+341.655746895" Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.274114 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.700891 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="293c612b-323d-4408-af28-b9f559766583" path="/var/lib/kubelet/pods/293c612b-323d-4408-af28-b9f559766583/volumes" Jan 20 08:55:07 crc kubenswrapper[4967]: I0120 08:55:07.701765 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac422dc4-1356-4cbe-b1ec-d99946de9a13" path="/var/lib/kubelet/pods/ac422dc4-1356-4cbe-b1ec-d99946de9a13/volumes" Jan 20 08:55:08 crc kubenswrapper[4967]: I0120 08:55:08.061586 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" event={"ID":"a429b03a-8a1d-479f-acd4-3aeebcc92d26","Type":"ContainerStarted","Data":"c22e35d87e3f3d5200165c5d59f60d8ac5ba0e8b807bf3b3c66764923a2a2ac6"} Jan 20 08:55:08 crc kubenswrapper[4967]: I0120 08:55:08.062013 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:08 crc kubenswrapper[4967]: I0120 08:55:08.066741 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:08 crc kubenswrapper[4967]: I0120 08:55:08.098154 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" podStartSLOduration=2.098131904 podStartE2EDuration="2.098131904s" podCreationTimestamp="2026-01-20 08:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:08.081815589 +0000 UTC m=+342.667115796" watchObservedRunningTime="2026-01-20 08:55:08.098131904 +0000 UTC m=+342.683432111" Jan 20 08:55:18 crc kubenswrapper[4967]: I0120 08:55:18.473952 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:55:18 crc kubenswrapper[4967]: I0120 08:55:18.474558 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.357141 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d2hw7"] Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.358808 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.376406 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d2hw7"] Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.550995 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-trusted-ca\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551041 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-certificates\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551075 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-bound-sa-token\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551092 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9czx\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-kube-api-access-m9czx\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551114 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-tls\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551141 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d99bef13-5f7e-4995-b186-907f9d20cda9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551348 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.551468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d99bef13-5f7e-4995-b186-907f9d20cda9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.573019 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.652929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9czx\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-kube-api-access-m9czx\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.652995 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-tls\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653045 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d99bef13-5f7e-4995-b186-907f9d20cda9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d99bef13-5f7e-4995-b186-907f9d20cda9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653164 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-trusted-ca\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653211 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-certificates\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653252 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-bound-sa-token\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.653876 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d99bef13-5f7e-4995-b186-907f9d20cda9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.654930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-certificates\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.655548 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d99bef13-5f7e-4995-b186-907f9d20cda9-trusted-ca\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.662583 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-registry-tls\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.665078 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d99bef13-5f7e-4995-b186-907f9d20cda9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.674909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9czx\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-kube-api-access-m9czx\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.679082 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d99bef13-5f7e-4995-b186-907f9d20cda9-bound-sa-token\") pod \"image-registry-66df7c8f76-d2hw7\" (UID: \"d99bef13-5f7e-4995-b186-907f9d20cda9\") " pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:44 crc kubenswrapper[4967]: I0120 08:55:44.976890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.025497 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.025771 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" podUID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" containerName="controller-manager" containerID="cri-o://c22e35d87e3f3d5200165c5d59f60d8ac5ba0e8b807bf3b3c66764923a2a2ac6" gracePeriod=30 Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.058278 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.058529 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" podUID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" containerName="route-controller-manager" containerID="cri-o://758b56e72e521fb7d3f82d064d1e52ff11e3ef394dbcb1e05cd7d3ecb7bcf3a0" gracePeriod=30 Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.302076 4967 generic.go:334] "Generic (PLEG): container finished" podID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" containerID="c22e35d87e3f3d5200165c5d59f60d8ac5ba0e8b807bf3b3c66764923a2a2ac6" exitCode=0 Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.302373 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" event={"ID":"a429b03a-8a1d-479f-acd4-3aeebcc92d26","Type":"ContainerDied","Data":"c22e35d87e3f3d5200165c5d59f60d8ac5ba0e8b807bf3b3c66764923a2a2ac6"} Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.304123 4967 generic.go:334] "Generic (PLEG): container finished" podID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" containerID="758b56e72e521fb7d3f82d064d1e52ff11e3ef394dbcb1e05cd7d3ecb7bcf3a0" exitCode=0 Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.304145 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" event={"ID":"e08d5d40-f3bd-446b-9687-1dc8420b92dd","Type":"ContainerDied","Data":"758b56e72e521fb7d3f82d064d1e52ff11e3ef394dbcb1e05cd7d3ecb7bcf3a0"} Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.483861 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.493643 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d2hw7"] Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.498604 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:45 crc kubenswrapper[4967]: W0120 08:55:45.499828 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd99bef13_5f7e_4995_b186_907f9d20cda9.slice/crio-ea9612148038feff1f63b6a5cc928ec201f26a99f9d8572d6f2a67d1ebbb14c9 WatchSource:0}: Error finding container ea9612148038feff1f63b6a5cc928ec201f26a99f9d8572d6f2a67d1ebbb14c9: Status 404 returned error can't find the container with id ea9612148038feff1f63b6a5cc928ec201f26a99f9d8572d6f2a67d1ebbb14c9 Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668386 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6htws\" (UniqueName: \"kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws\") pod \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668428 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config\") pod \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668452 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca\") pod \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668485 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config\") pod \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668511 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75gr\" (UniqueName: \"kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr\") pod \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668578 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert\") pod \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668647 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles\") pod \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668678 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca\") pod \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\" (UID: \"e08d5d40-f3bd-446b-9687-1dc8420b92dd\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.668710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert\") pod \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\" (UID: \"a429b03a-8a1d-479f-acd4-3aeebcc92d26\") " Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.669913 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config" (OuterVolumeSpecName: "config") pod "e08d5d40-f3bd-446b-9687-1dc8420b92dd" (UID: "e08d5d40-f3bd-446b-9687-1dc8420b92dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.670377 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca" (OuterVolumeSpecName: "client-ca") pod "a429b03a-8a1d-479f-acd4-3aeebcc92d26" (UID: "a429b03a-8a1d-479f-acd4-3aeebcc92d26"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.670448 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config" (OuterVolumeSpecName: "config") pod "a429b03a-8a1d-479f-acd4-3aeebcc92d26" (UID: "a429b03a-8a1d-479f-acd4-3aeebcc92d26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.670469 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a429b03a-8a1d-479f-acd4-3aeebcc92d26" (UID: "a429b03a-8a1d-479f-acd4-3aeebcc92d26"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.670924 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca" (OuterVolumeSpecName: "client-ca") pod "e08d5d40-f3bd-446b-9687-1dc8420b92dd" (UID: "e08d5d40-f3bd-446b-9687-1dc8420b92dd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.675642 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr" (OuterVolumeSpecName: "kube-api-access-r75gr") pod "a429b03a-8a1d-479f-acd4-3aeebcc92d26" (UID: "a429b03a-8a1d-479f-acd4-3aeebcc92d26"). InnerVolumeSpecName "kube-api-access-r75gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.676387 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws" (OuterVolumeSpecName: "kube-api-access-6htws") pod "e08d5d40-f3bd-446b-9687-1dc8420b92dd" (UID: "e08d5d40-f3bd-446b-9687-1dc8420b92dd"). InnerVolumeSpecName "kube-api-access-6htws". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.677554 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a429b03a-8a1d-479f-acd4-3aeebcc92d26" (UID: "a429b03a-8a1d-479f-acd4-3aeebcc92d26"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.689540 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e08d5d40-f3bd-446b-9687-1dc8420b92dd" (UID: "e08d5d40-f3bd-446b-9687-1dc8420b92dd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770755 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6htws\" (UniqueName: \"kubernetes.io/projected/e08d5d40-f3bd-446b-9687-1dc8420b92dd-kube-api-access-6htws\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770795 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770807 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770820 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-config\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770833 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75gr\" (UniqueName: \"kubernetes.io/projected/a429b03a-8a1d-479f-acd4-3aeebcc92d26-kube-api-access-r75gr\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770842 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08d5d40-f3bd-446b-9687-1dc8420b92dd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770850 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a429b03a-8a1d-479f-acd4-3aeebcc92d26-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770858 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e08d5d40-f3bd-446b-9687-1dc8420b92dd-client-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:45 crc kubenswrapper[4967]: I0120 08:55:45.770866 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a429b03a-8a1d-479f-acd4-3aeebcc92d26-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.312272 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" event={"ID":"a429b03a-8a1d-479f-acd4-3aeebcc92d26","Type":"ContainerDied","Data":"659148c3303728e54094f53b2cea6c735c2a3e6e2a8f57b6a8442f3b11731f8a"} Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.312327 4967 scope.go:117] "RemoveContainer" containerID="c22e35d87e3f3d5200165c5d59f60d8ac5ba0e8b807bf3b3c66764923a2a2ac6" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.312522 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.314797 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" event={"ID":"d99bef13-5f7e-4995-b186-907f9d20cda9","Type":"ContainerStarted","Data":"ade389b727013680c9d6bdee9c2c19653071c3311c41d1cdbaeeafbbb8e41f45"} Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.314846 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" event={"ID":"d99bef13-5f7e-4995-b186-907f9d20cda9","Type":"ContainerStarted","Data":"ea9612148038feff1f63b6a5cc928ec201f26a99f9d8572d6f2a67d1ebbb14c9"} Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.314892 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.315917 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.316304 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9pxql" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="registry-server" containerID="cri-o://f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307" gracePeriod=2 Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.316644 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" event={"ID":"e08d5d40-f3bd-446b-9687-1dc8420b92dd","Type":"ContainerDied","Data":"2e71f1496de1eaf6214dff57a8c1620fe44068592fb28845d2ec13e97db83781"} Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.316778 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.336970 4967 scope.go:117] "RemoveContainer" containerID="758b56e72e521fb7d3f82d064d1e52ff11e3ef394dbcb1e05cd7d3ecb7bcf3a0" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.364151 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" podStartSLOduration=2.363480474 podStartE2EDuration="2.363480474s" podCreationTimestamp="2026-01-20 08:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:46.35896127 +0000 UTC m=+380.944261507" watchObservedRunningTime="2026-01-20 08:55:46.363480474 +0000 UTC m=+380.948780761" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.377505 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.381204 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5ff6ccf94b-nmhfl"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.393645 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.397977 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dfff9545c-nfltr"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.521054 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.521412 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pn7jw" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="registry-server" containerID="cri-o://af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41" gracePeriod=2 Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.674779 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.791626 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jslhp\" (UniqueName: \"kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp\") pod \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.791749 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities\") pod \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.791778 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content\") pod \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\" (UID: \"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.792496 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities" (OuterVolumeSpecName: "utilities") pod "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" (UID: "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.792845 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.797053 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp" (OuterVolumeSpecName: "kube-api-access-jslhp") pod "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" (UID: "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e"). InnerVolumeSpecName "kube-api-access-jslhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.815354 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" (UID: "f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.895385 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.895422 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jslhp\" (UniqueName: \"kubernetes.io/projected/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e-kube-api-access-jslhp\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.901885 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.996011 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content\") pod \"8b655a5d-4022-48f0-a80e-e790a9235b3c\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.996126 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5x59\" (UniqueName: \"kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59\") pod \"8b655a5d-4022-48f0-a80e-e790a9235b3c\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.996211 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities\") pod \"8b655a5d-4022-48f0-a80e-e790a9235b3c\" (UID: \"8b655a5d-4022-48f0-a80e-e790a9235b3c\") " Jan 20 08:55:46 crc kubenswrapper[4967]: I0120 08:55:46.999672 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities" (OuterVolumeSpecName: "utilities") pod "8b655a5d-4022-48f0-a80e-e790a9235b3c" (UID: "8b655a5d-4022-48f0-a80e-e790a9235b3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.001670 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59" (OuterVolumeSpecName: "kube-api-access-m5x59") pod "8b655a5d-4022-48f0-a80e-e790a9235b3c" (UID: "8b655a5d-4022-48f0-a80e-e790a9235b3c"). InnerVolumeSpecName "kube-api-access-m5x59". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087447 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz"] Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087797 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="extract-utilities" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087819 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="extract-utilities" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087837 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="extract-content" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087845 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="extract-content" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087859 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" containerName="controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087869 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" containerName="controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087888 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="extract-utilities" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087895 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="extract-utilities" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087907 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="extract-content" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087914 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="extract-content" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087930 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087938 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087950 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" containerName="route-controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087958 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" containerName="route-controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.087969 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.087979 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.088105 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.088122 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" containerName="controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.088134 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerName="registry-server" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.088149 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" containerName="route-controller-manager" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.088732 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.091054 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.091867 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.092165 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.092247 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.092390 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.092993 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.093234 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.093450 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.093971 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.095677 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.095739 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.095941 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.097016 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.097946 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.097999 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5x59\" (UniqueName: \"kubernetes.io/projected/8b655a5d-4022-48f0-a80e-e790a9235b3c-kube-api-access-m5x59\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.101007 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.101134 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.102943 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.119299 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.144025 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b655a5d-4022-48f0-a80e-e790a9235b3c" (UID: "8b655a5d-4022-48f0-a80e-e790a9235b3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.198985 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-proxy-ca-bundles\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199028 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d86efe98-71b7-489f-9f59-e0be47360898-serving-cert\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199063 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-config\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199080 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-client-ca\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199097 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-config\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199116 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fc36466-d8c6-485b-940d-d368e661d0ba-serving-cert\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199132 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm2xn\" (UniqueName: \"kubernetes.io/projected/4fc36466-d8c6-485b-940d-d368e661d0ba-kube-api-access-gm2xn\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199155 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq6ld\" (UniqueName: \"kubernetes.io/projected/d86efe98-71b7-489f-9f59-e0be47360898-kube-api-access-tq6ld\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199187 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-client-ca\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.199219 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b655a5d-4022-48f0-a80e-e790a9235b3c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300090 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-proxy-ca-bundles\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300153 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d86efe98-71b7-489f-9f59-e0be47360898-serving-cert\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300206 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-config\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300228 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-client-ca\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-config\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300276 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fc36466-d8c6-485b-940d-d368e661d0ba-serving-cert\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300297 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm2xn\" (UniqueName: \"kubernetes.io/projected/4fc36466-d8c6-485b-940d-d368e661d0ba-kube-api-access-gm2xn\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300326 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq6ld\" (UniqueName: \"kubernetes.io/projected/d86efe98-71b7-489f-9f59-e0be47360898-kube-api-access-tq6ld\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.300370 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-client-ca\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.301464 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-client-ca\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.302288 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-proxy-ca-bundles\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.302733 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-config\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.304063 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4fc36466-d8c6-485b-940d-d368e661d0ba-client-ca\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.305375 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d86efe98-71b7-489f-9f59-e0be47360898-serving-cert\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.305896 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d86efe98-71b7-489f-9f59-e0be47360898-config\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.307361 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fc36466-d8c6-485b-940d-d368e661d0ba-serving-cert\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.318352 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm2xn\" (UniqueName: \"kubernetes.io/projected/4fc36466-d8c6-485b-940d-d368e661d0ba-kube-api-access-gm2xn\") pod \"controller-manager-75c5d6d4b9-gvhs4\" (UID: \"4fc36466-d8c6-485b-940d-d368e661d0ba\") " pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.320231 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq6ld\" (UniqueName: \"kubernetes.io/projected/d86efe98-71b7-489f-9f59-e0be47360898-kube-api-access-tq6ld\") pod \"route-controller-manager-6bf9b6884f-96gdz\" (UID: \"d86efe98-71b7-489f-9f59-e0be47360898\") " pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.323752 4967 generic.go:334] "Generic (PLEG): container finished" podID="8b655a5d-4022-48f0-a80e-e790a9235b3c" containerID="af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41" exitCode=0 Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.323841 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn7jw" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.323853 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerDied","Data":"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41"} Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.323891 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn7jw" event={"ID":"8b655a5d-4022-48f0-a80e-e790a9235b3c","Type":"ContainerDied","Data":"f2f0670a05d483844edd96e9772965c58d9fd28322eb4fc719e08f72f490d1be"} Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.323922 4967 scope.go:117] "RemoveContainer" containerID="af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.329416 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" containerID="f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307" exitCode=0 Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.329469 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerDied","Data":"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307"} Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.329512 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pxql" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.329546 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pxql" event={"ID":"f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e","Type":"ContainerDied","Data":"823b1349e871568e622694b8fa70f0785151b474687aad2e72f5d00429c5c88d"} Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.337123 4967 scope.go:117] "RemoveContainer" containerID="55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.351809 4967 scope.go:117] "RemoveContainer" containerID="aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.375932 4967 scope.go:117] "RemoveContainer" containerID="af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.376716 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41\": container with ID starting with af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41 not found: ID does not exist" containerID="af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.376778 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41"} err="failed to get container status \"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41\": rpc error: code = NotFound desc = could not find container \"af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41\": container with ID starting with af4652328b6b7d31dde2537f8ba8e9bd77af4954c0f3e1540c4cd461f3f8ba41 not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.376819 4967 scope.go:117] "RemoveContainer" containerID="55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.377050 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.377185 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6\": container with ID starting with 55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6 not found: ID does not exist" containerID="55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.377228 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6"} err="failed to get container status \"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6\": rpc error: code = NotFound desc = could not find container \"55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6\": container with ID starting with 55b2a011b94e6dffb01c04029a797b8eb54af69638db4aa6be39c72d02482dd6 not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.377254 4967 scope.go:117] "RemoveContainer" containerID="aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.377510 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9\": container with ID starting with aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9 not found: ID does not exist" containerID="aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.378810 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9"} err="failed to get container status \"aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9\": rpc error: code = NotFound desc = could not find container \"aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9\": container with ID starting with aac98524e145f3755cfc22b1505bf1219dced27fed38ef88587fff03e9ed0ca9 not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.378872 4967 scope.go:117] "RemoveContainer" containerID="f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.380794 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pn7jw"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.393177 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.397877 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pxql"] Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.398332 4967 scope.go:117] "RemoveContainer" containerID="20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.412107 4967 scope.go:117] "RemoveContainer" containerID="b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.414892 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.423990 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.439403 4967 scope.go:117] "RemoveContainer" containerID="f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.440192 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307\": container with ID starting with f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307 not found: ID does not exist" containerID="f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.440222 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307"} err="failed to get container status \"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307\": rpc error: code = NotFound desc = could not find container \"f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307\": container with ID starting with f6333ac939e8fed4dce3a42093b7800c4d4051df57bc86c7d19aa32b44718307 not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.440243 4967 scope.go:117] "RemoveContainer" containerID="20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.440541 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4\": container with ID starting with 20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4 not found: ID does not exist" containerID="20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.440578 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4"} err="failed to get container status \"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4\": rpc error: code = NotFound desc = could not find container \"20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4\": container with ID starting with 20df39a58a0f6086626b81e65569d42a2d8a870300e8f5e5b8697be08744aba4 not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.440604 4967 scope.go:117] "RemoveContainer" containerID="b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c" Jan 20 08:55:47 crc kubenswrapper[4967]: E0120 08:55:47.440859 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c\": container with ID starting with b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c not found: ID does not exist" containerID="b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.441008 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c"} err="failed to get container status \"b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c\": rpc error: code = NotFound desc = could not find container \"b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c\": container with ID starting with b609c44dce1dde03ca5b6a3de6dd300ae6192da5dc52c7b41b4a58c1a74a937c not found: ID does not exist" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.623940 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz"] Jan 20 08:55:47 crc kubenswrapper[4967]: W0120 08:55:47.639993 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd86efe98_71b7_489f_9f59_e0be47360898.slice/crio-52b022ddba4a4c75dd59819531f635784451d2bb2650c59a81fb97e5af2c3736 WatchSource:0}: Error finding container 52b022ddba4a4c75dd59819531f635784451d2bb2650c59a81fb97e5af2c3736: Status 404 returned error can't find the container with id 52b022ddba4a4c75dd59819531f635784451d2bb2650c59a81fb97e5af2c3736 Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.701931 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b655a5d-4022-48f0-a80e-e790a9235b3c" path="/var/lib/kubelet/pods/8b655a5d-4022-48f0-a80e-e790a9235b3c/volumes" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.703527 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a429b03a-8a1d-479f-acd4-3aeebcc92d26" path="/var/lib/kubelet/pods/a429b03a-8a1d-479f-acd4-3aeebcc92d26/volumes" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.704260 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e08d5d40-f3bd-446b-9687-1dc8420b92dd" path="/var/lib/kubelet/pods/e08d5d40-f3bd-446b-9687-1dc8420b92dd/volumes" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.705460 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e" path="/var/lib/kubelet/pods/f0b4d42d-49d9-43ac-b60d-e4a9b10f7f4e/volumes" Jan 20 08:55:47 crc kubenswrapper[4967]: I0120 08:55:47.875728 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4"] Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.337724 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" event={"ID":"4fc36466-d8c6-485b-940d-d368e661d0ba","Type":"ContainerStarted","Data":"3cf56dfe60281432232fa3df467442fa74fadc6e51fba30386dd9ef0c1ed424e"} Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.337780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" event={"ID":"4fc36466-d8c6-485b-940d-d368e661d0ba","Type":"ContainerStarted","Data":"40c1d959bf3afaa4054ec3fb06ee1eec2d0a3d12aa90130722008b34506d73b9"} Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.338009 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.342168 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" event={"ID":"d86efe98-71b7-489f-9f59-e0be47360898","Type":"ContainerStarted","Data":"cc12192f2386bf8d82765d0050c1affa8217efd64771aea1c11dd3c566d350d4"} Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.342213 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" event={"ID":"d86efe98-71b7-489f-9f59-e0be47360898","Type":"ContainerStarted","Data":"52b022ddba4a4c75dd59819531f635784451d2bb2650c59a81fb97e5af2c3736"} Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.342392 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.344662 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.348221 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.361803 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75c5d6d4b9-gvhs4" podStartSLOduration=3.36177503 podStartE2EDuration="3.36177503s" podCreationTimestamp="2026-01-20 08:55:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:48.356458405 +0000 UTC m=+382.941758622" watchObservedRunningTime="2026-01-20 08:55:48.36177503 +0000 UTC m=+382.947075237" Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.474130 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:55:48 crc kubenswrapper[4967]: I0120 08:55:48.474185 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.491073 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6bf9b6884f-96gdz" podStartSLOduration=4.491047236 podStartE2EDuration="4.491047236s" podCreationTimestamp="2026-01-20 08:55:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:48.390855963 +0000 UTC m=+382.976156170" watchObservedRunningTime="2026-01-20 08:55:49.491047236 +0000 UTC m=+384.076347443" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.496298 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.496718 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qmfmw" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="registry-server" containerID="cri-o://63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0" gracePeriod=30 Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.502885 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.503651 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jbrxm" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="registry-server" containerID="cri-o://03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205" gracePeriod=30 Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.509815 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.510208 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" containerID="cri-o://0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275" gracePeriod=30 Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.528279 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.528632 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d9b7s" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" containerID="cri-o://70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" gracePeriod=30 Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.543875 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqc5w"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.546274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.552195 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqc5w"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.574750 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.575025 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2lz9v" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="registry-server" containerID="cri-o://2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416" gracePeriod=30 Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.644005 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.644092 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.644121 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt2k9\" (UniqueName: \"kubernetes.io/projected/778bd2b6-c30e-43eb-8045-ff0720482f5c-kube-api-access-rt2k9\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.745364 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.745694 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt2k9\" (UniqueName: \"kubernetes.io/projected/778bd2b6-c30e-43eb-8045-ff0720482f5c-kube-api-access-rt2k9\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.745757 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.746961 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.752113 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/778bd2b6-c30e-43eb-8045-ff0720482f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.763096 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt2k9\" (UniqueName: \"kubernetes.io/projected/778bd2b6-c30e-43eb-8045-ff0720482f5c-kube-api-access-rt2k9\") pod \"marketplace-operator-79b997595-sqc5w\" (UID: \"778bd2b6-c30e-43eb-8045-ff0720482f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: E0120 08:55:49.794338 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 is running failed: container process not found" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:55:49 crc kubenswrapper[4967]: E0120 08:55:49.795171 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 is running failed: container process not found" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:55:49 crc kubenswrapper[4967]: E0120 08:55:49.796202 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 is running failed: container process not found" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 08:55:49 crc kubenswrapper[4967]: E0120 08:55:49.796238 4967 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-d9b7s" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.977736 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:49 crc kubenswrapper[4967]: I0120 08:55:49.993396 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.035098 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.036783 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.044340 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.066309 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153167 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvtd8\" (UniqueName: \"kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8\") pod \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153473 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf42b\" (UniqueName: \"kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b\") pod \"588d30c5-1a85-42cf-8b19-e12755604f91\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153501 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca\") pod \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content\") pod \"588d30c5-1a85-42cf-8b19-e12755604f91\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153579 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content\") pod \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153623 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities\") pod \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153660 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities\") pod \"588d30c5-1a85-42cf-8b19-e12755604f91\" (UID: \"588d30c5-1a85-42cf-8b19-e12755604f91\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153679 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdp77\" (UniqueName: \"kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77\") pod \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\" (UID: \"5216e81e-0c81-4fc3-9fa2-458e24ed883d\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153697 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities\") pod \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153721 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities\") pod \"551f69d7-2c12-4474-9169-4540933734e1\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153741 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content\") pod \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153759 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics\") pod \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\" (UID: \"defe6cb5-e637-4f50-a67a-7d1c6ae28879\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153780 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl4zq\" (UniqueName: \"kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq\") pod \"551f69d7-2c12-4474-9169-4540933734e1\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153813 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content\") pod \"551f69d7-2c12-4474-9169-4540933734e1\" (UID: \"551f69d7-2c12-4474-9169-4540933734e1\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.153829 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kwcf\" (UniqueName: \"kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf\") pod \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\" (UID: \"a0eaef46-8b17-45f6-9786-da4ebff48dfe\") " Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.154418 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "defe6cb5-e637-4f50-a67a-7d1c6ae28879" (UID: "defe6cb5-e637-4f50-a67a-7d1c6ae28879"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.155021 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities" (OuterVolumeSpecName: "utilities") pod "a0eaef46-8b17-45f6-9786-da4ebff48dfe" (UID: "a0eaef46-8b17-45f6-9786-da4ebff48dfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.155886 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities" (OuterVolumeSpecName: "utilities") pod "551f69d7-2c12-4474-9169-4540933734e1" (UID: "551f69d7-2c12-4474-9169-4540933734e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.159256 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities" (OuterVolumeSpecName: "utilities") pod "5216e81e-0c81-4fc3-9fa2-458e24ed883d" (UID: "5216e81e-0c81-4fc3-9fa2-458e24ed883d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.159451 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq" (OuterVolumeSpecName: "kube-api-access-wl4zq") pod "551f69d7-2c12-4474-9169-4540933734e1" (UID: "551f69d7-2c12-4474-9169-4540933734e1"). InnerVolumeSpecName "kube-api-access-wl4zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.159655 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities" (OuterVolumeSpecName: "utilities") pod "588d30c5-1a85-42cf-8b19-e12755604f91" (UID: "588d30c5-1a85-42cf-8b19-e12755604f91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.159862 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "defe6cb5-e637-4f50-a67a-7d1c6ae28879" (UID: "defe6cb5-e637-4f50-a67a-7d1c6ae28879"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.160512 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b" (OuterVolumeSpecName: "kube-api-access-rf42b") pod "588d30c5-1a85-42cf-8b19-e12755604f91" (UID: "588d30c5-1a85-42cf-8b19-e12755604f91"). InnerVolumeSpecName "kube-api-access-rf42b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.160574 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8" (OuterVolumeSpecName: "kube-api-access-kvtd8") pod "defe6cb5-e637-4f50-a67a-7d1c6ae28879" (UID: "defe6cb5-e637-4f50-a67a-7d1c6ae28879"). InnerVolumeSpecName "kube-api-access-kvtd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.161591 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77" (OuterVolumeSpecName: "kube-api-access-mdp77") pod "5216e81e-0c81-4fc3-9fa2-458e24ed883d" (UID: "5216e81e-0c81-4fc3-9fa2-458e24ed883d"). InnerVolumeSpecName "kube-api-access-mdp77". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.180892 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf" (OuterVolumeSpecName: "kube-api-access-8kwcf") pod "a0eaef46-8b17-45f6-9786-da4ebff48dfe" (UID: "a0eaef46-8b17-45f6-9786-da4ebff48dfe"). InnerVolumeSpecName "kube-api-access-8kwcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.190215 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "551f69d7-2c12-4474-9169-4540933734e1" (UID: "551f69d7-2c12-4474-9169-4540933734e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.217748 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5216e81e-0c81-4fc3-9fa2-458e24ed883d" (UID: "5216e81e-0c81-4fc3-9fa2-458e24ed883d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.221462 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0eaef46-8b17-45f6-9786-da4ebff48dfe" (UID: "a0eaef46-8b17-45f6-9786-da4ebff48dfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255694 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255731 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5216e81e-0c81-4fc3-9fa2-458e24ed883d-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255743 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255755 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdp77\" (UniqueName: \"kubernetes.io/projected/5216e81e-0c81-4fc3-9fa2-458e24ed883d-kube-api-access-mdp77\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255771 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255782 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255792 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eaef46-8b17-45f6-9786-da4ebff48dfe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255803 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255816 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl4zq\" (UniqueName: \"kubernetes.io/projected/551f69d7-2c12-4474-9169-4540933734e1-kube-api-access-wl4zq\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255827 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/551f69d7-2c12-4474-9169-4540933734e1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255839 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kwcf\" (UniqueName: \"kubernetes.io/projected/a0eaef46-8b17-45f6-9786-da4ebff48dfe-kube-api-access-8kwcf\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255850 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvtd8\" (UniqueName: \"kubernetes.io/projected/defe6cb5-e637-4f50-a67a-7d1c6ae28879-kube-api-access-kvtd8\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255862 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf42b\" (UniqueName: \"kubernetes.io/projected/588d30c5-1a85-42cf-8b19-e12755604f91-kube-api-access-rf42b\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.255874 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/defe6cb5-e637-4f50-a67a-7d1c6ae28879-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.276210 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "588d30c5-1a85-42cf-8b19-e12755604f91" (UID: "588d30c5-1a85-42cf-8b19-e12755604f91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356522 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerID="03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205" exitCode=0 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356556 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d30c5-1a85-42cf-8b19-e12755604f91-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356580 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerDied","Data":"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356620 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbrxm" event={"ID":"a0eaef46-8b17-45f6-9786-da4ebff48dfe","Type":"ContainerDied","Data":"042919757c6c32f23230cde2713f81b51c2846cb94fe329e40327c2bd522af21"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356637 4967 scope.go:117] "RemoveContainer" containerID="03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.356681 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbrxm" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.365625 4967 generic.go:334] "Generic (PLEG): container finished" podID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerID="63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0" exitCode=0 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.365684 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerDied","Data":"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.365709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmfmw" event={"ID":"5216e81e-0c81-4fc3-9fa2-458e24ed883d","Type":"ContainerDied","Data":"162c24c565b61a43160e5a0cd43eac9ef917b28adccaebc682d91d9eace78da6"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.365787 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmfmw" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.374392 4967 generic.go:334] "Generic (PLEG): container finished" podID="551f69d7-2c12-4474-9169-4540933734e1" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" exitCode=0 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.374491 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerDied","Data":"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.374522 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9b7s" event={"ID":"551f69d7-2c12-4474-9169-4540933734e1","Type":"ContainerDied","Data":"d910483add5080a8ec3f3557c7c322e06a18b724baa77a625390b4607aecff01"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.374604 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9b7s" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.375346 4967 scope.go:117] "RemoveContainer" containerID="c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.376787 4967 generic.go:334] "Generic (PLEG): container finished" podID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerID="0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275" exitCode=0 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.376840 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerDied","Data":"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.376867 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" event={"ID":"defe6cb5-e637-4f50-a67a-7d1c6ae28879","Type":"ContainerDied","Data":"d547ae833f1c3604779048bda3227dc881f73c5ca28611cfa0837ed751bffb83"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.376919 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksh2s" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.379783 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lz9v" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.379801 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerDied","Data":"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.379734 4967 generic.go:334] "Generic (PLEG): container finished" podID="588d30c5-1a85-42cf-8b19-e12755604f91" containerID="2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416" exitCode=0 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.380077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lz9v" event={"ID":"588d30c5-1a85-42cf-8b19-e12755604f91","Type":"ContainerDied","Data":"15b2fe3c025d75936fb3ce7325ee2bc61a83857717c415b832a46b461a205511"} Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.389510 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.394703 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jbrxm"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.417197 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.419998 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqc5w"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.422649 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksh2s"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.422893 4967 scope.go:117] "RemoveContainer" containerID="17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3" Jan 20 08:55:50 crc kubenswrapper[4967]: W0120 08:55:50.433902 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod778bd2b6_c30e_43eb_8045_ff0720482f5c.slice/crio-64ecaeedd4f0abada575af3503cba03b992b9cfd317508079c9a8c865aeb65b1 WatchSource:0}: Error finding container 64ecaeedd4f0abada575af3503cba03b992b9cfd317508079c9a8c865aeb65b1: Status 404 returned error can't find the container with id 64ecaeedd4f0abada575af3503cba03b992b9cfd317508079c9a8c865aeb65b1 Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.440641 4967 scope.go:117] "RemoveContainer" containerID="03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.443411 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.446185 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205\": container with ID starting with 03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205 not found: ID does not exist" containerID="03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.446224 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205"} err="failed to get container status \"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205\": rpc error: code = NotFound desc = could not find container \"03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205\": container with ID starting with 03f9e9be70feafd7593e56409368576f3045adb8739f759c39d61f1c1bbab205 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.446260 4967 scope.go:117] "RemoveContainer" containerID="c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.446829 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e\": container with ID starting with c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e not found: ID does not exist" containerID="c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.446860 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e"} err="failed to get container status \"c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e\": rpc error: code = NotFound desc = could not find container \"c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e\": container with ID starting with c403eea553408eb940f49cfb157ed978695d8c8d4ab8c20b4254d3b1d934af1e not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.446883 4967 scope.go:117] "RemoveContainer" containerID="17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.447180 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3\": container with ID starting with 17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3 not found: ID does not exist" containerID="17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.447204 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3"} err="failed to get container status \"17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3\": rpc error: code = NotFound desc = could not find container \"17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3\": container with ID starting with 17645c3d172825d807db8c2f7cd85c361ae6e5c43b34a1c845428bfb5e6127c3 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.447221 4967 scope.go:117] "RemoveContainer" containerID="63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.458567 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9b7s"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.467178 4967 scope.go:117] "RemoveContainer" containerID="a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.468211 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.472401 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2lz9v"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.477262 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.479417 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qmfmw"] Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.480822 4967 scope.go:117] "RemoveContainer" containerID="3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.501907 4967 scope.go:117] "RemoveContainer" containerID="63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.502352 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0\": container with ID starting with 63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0 not found: ID does not exist" containerID="63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.502383 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0"} err="failed to get container status \"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0\": rpc error: code = NotFound desc = could not find container \"63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0\": container with ID starting with 63304a60b549f9133322257ae5d3a27adda62a32804ac4a27e0439cc03df1db0 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.502409 4967 scope.go:117] "RemoveContainer" containerID="a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.503602 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a\": container with ID starting with a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a not found: ID does not exist" containerID="a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.503656 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a"} err="failed to get container status \"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a\": rpc error: code = NotFound desc = could not find container \"a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a\": container with ID starting with a5bb857b3fb102ec69efed83035acece960f1dd970835c760effe2460ff7cc0a not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.503679 4967 scope.go:117] "RemoveContainer" containerID="3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.504249 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c\": container with ID starting with 3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c not found: ID does not exist" containerID="3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.504289 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c"} err="failed to get container status \"3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c\": rpc error: code = NotFound desc = could not find container \"3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c\": container with ID starting with 3ea873ca9d076e712568bb09d0dd44b69bf1e51df2c541eefdafca33b86dc98c not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.504316 4967 scope.go:117] "RemoveContainer" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.531571 4967 scope.go:117] "RemoveContainer" containerID="ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.552999 4967 scope.go:117] "RemoveContainer" containerID="38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.568078 4967 scope.go:117] "RemoveContainer" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.568560 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9\": container with ID starting with 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 not found: ID does not exist" containerID="70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.568598 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9"} err="failed to get container status \"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9\": rpc error: code = NotFound desc = could not find container \"70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9\": container with ID starting with 70bb348fd15afddf7898c0a8add41894dfbd4a74d597fd9710985bfbc50abea9 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.568641 4967 scope.go:117] "RemoveContainer" containerID="ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.568917 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1\": container with ID starting with ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1 not found: ID does not exist" containerID="ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.568939 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1"} err="failed to get container status \"ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1\": rpc error: code = NotFound desc = could not find container \"ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1\": container with ID starting with ce4eae8cc86e61742b38e86d74602828901879513b46b31e3e14ea331c68f6e1 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.568953 4967 scope.go:117] "RemoveContainer" containerID="38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.569302 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e\": container with ID starting with 38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e not found: ID does not exist" containerID="38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.569364 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e"} err="failed to get container status \"38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e\": rpc error: code = NotFound desc = could not find container \"38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e\": container with ID starting with 38e98d5faa7792efcca4d4bae7aa8b4b05751617222ccef22d525b45b6b09e3e not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.569400 4967 scope.go:117] "RemoveContainer" containerID="0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.582840 4967 scope.go:117] "RemoveContainer" containerID="fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.596654 4967 scope.go:117] "RemoveContainer" containerID="0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.597326 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275\": container with ID starting with 0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275 not found: ID does not exist" containerID="0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.597394 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275"} err="failed to get container status \"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275\": rpc error: code = NotFound desc = could not find container \"0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275\": container with ID starting with 0dd574a746a3c7bad1fa4141c448f89e7c4b5721a17ce6409451ac4fa101b275 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.597435 4967 scope.go:117] "RemoveContainer" containerID="fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.597827 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11\": container with ID starting with fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11 not found: ID does not exist" containerID="fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.597864 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11"} err="failed to get container status \"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11\": rpc error: code = NotFound desc = could not find container \"fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11\": container with ID starting with fa201bc88b3c801695798130aecb396b00b3e98133dc0d47283ca8365962fd11 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.597888 4967 scope.go:117] "RemoveContainer" containerID="2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.613786 4967 scope.go:117] "RemoveContainer" containerID="83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.661462 4967 scope.go:117] "RemoveContainer" containerID="6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.676127 4967 scope.go:117] "RemoveContainer" containerID="2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.676685 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416\": container with ID starting with 2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416 not found: ID does not exist" containerID="2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.676712 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416"} err="failed to get container status \"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416\": rpc error: code = NotFound desc = could not find container \"2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416\": container with ID starting with 2569e143a039e9e3f3944a72cd1ea8b99baad01944ef6071efd225b5e9606416 not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.676734 4967 scope.go:117] "RemoveContainer" containerID="83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.677110 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b\": container with ID starting with 83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b not found: ID does not exist" containerID="83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.677131 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b"} err="failed to get container status \"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b\": rpc error: code = NotFound desc = could not find container \"83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b\": container with ID starting with 83091d0e877b8470da5be19ab55e511e68d16559e7ab600ce4514b1fac16a89b not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.677145 4967 scope.go:117] "RemoveContainer" containerID="6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.677511 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a\": container with ID starting with 6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a not found: ID does not exist" containerID="6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.677548 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a"} err="failed to get container status \"6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a\": rpc error: code = NotFound desc = could not find container \"6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a\": container with ID starting with 6c906a38e0e7f770a3a7ff30944c8da97f63f61ae1640bffb00eb6707894a13a not found: ID does not exist" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.920750 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q6d4w"] Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921032 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921056 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921070 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921080 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921104 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921114 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921124 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921132 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921146 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921154 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921166 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921174 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="extract-utilities" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921188 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921197 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921209 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921218 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921230 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921238 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921250 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921258 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921268 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921277 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921289 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921314 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921323 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921330 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: E0120 08:55:50.921341 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921349 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="extract-content" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921466 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921484 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921501 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921511 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="551f69d7-2c12-4474-9169-4540933734e1" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921522 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" containerName="registry-server" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.921746 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" containerName="marketplace-operator" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.922415 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.925008 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 20 08:55:50 crc kubenswrapper[4967]: I0120 08:55:50.973807 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q6d4w"] Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.076917 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-utilities\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.077018 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-catalog-content\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.077075 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfgdx\" (UniqueName: \"kubernetes.io/projected/b1cfca35-a47d-4fc4-a425-57a60dd4b855-kube-api-access-sfgdx\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.178599 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-catalog-content\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.178763 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfgdx\" (UniqueName: \"kubernetes.io/projected/b1cfca35-a47d-4fc4-a425-57a60dd4b855-kube-api-access-sfgdx\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.178796 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-utilities\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.179456 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-catalog-content\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.179721 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1cfca35-a47d-4fc4-a425-57a60dd4b855-utilities\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.230543 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfgdx\" (UniqueName: \"kubernetes.io/projected/b1cfca35-a47d-4fc4-a425-57a60dd4b855-kube-api-access-sfgdx\") pod \"redhat-operators-q6d4w\" (UID: \"b1cfca35-a47d-4fc4-a425-57a60dd4b855\") " pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.303619 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.387175 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" event={"ID":"778bd2b6-c30e-43eb-8045-ff0720482f5c","Type":"ContainerStarted","Data":"0c5aee72081c856a04dde33c86a352ea806cf20fb4c60a0f7b3591d09ebaa15f"} Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.388308 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" event={"ID":"778bd2b6-c30e-43eb-8045-ff0720482f5c","Type":"ContainerStarted","Data":"64ecaeedd4f0abada575af3503cba03b992b9cfd317508079c9a8c865aeb65b1"} Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.389627 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.399322 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.430976 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sqc5w" podStartSLOduration=2.430932931 podStartE2EDuration="2.430932931s" podCreationTimestamp="2026-01-20 08:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 08:55:51.409213189 +0000 UTC m=+385.994513416" watchObservedRunningTime="2026-01-20 08:55:51.430932931 +0000 UTC m=+386.016233138" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.703602 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5216e81e-0c81-4fc3-9fa2-458e24ed883d" path="/var/lib/kubelet/pods/5216e81e-0c81-4fc3-9fa2-458e24ed883d/volumes" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.705778 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="551f69d7-2c12-4474-9169-4540933734e1" path="/var/lib/kubelet/pods/551f69d7-2c12-4474-9169-4540933734e1/volumes" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.706567 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="588d30c5-1a85-42cf-8b19-e12755604f91" path="/var/lib/kubelet/pods/588d30c5-1a85-42cf-8b19-e12755604f91/volumes" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.707284 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0eaef46-8b17-45f6-9786-da4ebff48dfe" path="/var/lib/kubelet/pods/a0eaef46-8b17-45f6-9786-da4ebff48dfe/volumes" Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.708534 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="defe6cb5-e637-4f50-a67a-7d1c6ae28879" path="/var/lib/kubelet/pods/defe6cb5-e637-4f50-a67a-7d1c6ae28879/volumes" Jan 20 08:55:51 crc kubenswrapper[4967]: W0120 08:55:51.709001 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1cfca35_a47d_4fc4_a425_57a60dd4b855.slice/crio-e97b161d764b9584ed4f965e63eaffc3da421e801c4c8a42ca5716dffc89c388 WatchSource:0}: Error finding container e97b161d764b9584ed4f965e63eaffc3da421e801c4c8a42ca5716dffc89c388: Status 404 returned error can't find the container with id e97b161d764b9584ed4f965e63eaffc3da421e801c4c8a42ca5716dffc89c388 Jan 20 08:55:51 crc kubenswrapper[4967]: I0120 08:55:51.709030 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q6d4w"] Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.404192 4967 generic.go:334] "Generic (PLEG): container finished" podID="b1cfca35-a47d-4fc4-a425-57a60dd4b855" containerID="efc82140a204bda87ae1ea515478cfc4955f17f2f895c6a7e20968edccb29171" exitCode=0 Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.404291 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q6d4w" event={"ID":"b1cfca35-a47d-4fc4-a425-57a60dd4b855","Type":"ContainerDied","Data":"efc82140a204bda87ae1ea515478cfc4955f17f2f895c6a7e20968edccb29171"} Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.405791 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q6d4w" event={"ID":"b1cfca35-a47d-4fc4-a425-57a60dd4b855","Type":"ContainerStarted","Data":"e97b161d764b9584ed4f965e63eaffc3da421e801c4c8a42ca5716dffc89c388"} Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.718793 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.719735 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.725804 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.730345 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.799254 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.799495 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.799543 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrpr\" (UniqueName: \"kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.900385 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.900422 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrpr\" (UniqueName: \"kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.900490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.900931 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.900939 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:52 crc kubenswrapper[4967]: I0120 08:55:52.918941 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrpr\" (UniqueName: \"kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr\") pod \"certified-operators-s5trl\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.042946 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.320643 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xhnv8"] Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.321820 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.327392 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.332629 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xhnv8"] Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.409281 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz6k2\" (UniqueName: \"kubernetes.io/projected/a2774890-f4e7-47b2-97c2-d308e78b7256-kube-api-access-tz6k2\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.409345 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-catalog-content\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.409481 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-utilities\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.447106 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.510900 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz6k2\" (UniqueName: \"kubernetes.io/projected/a2774890-f4e7-47b2-97c2-d308e78b7256-kube-api-access-tz6k2\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.510956 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-catalog-content\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.510990 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-utilities\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.511354 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-utilities\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.512827 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2774890-f4e7-47b2-97c2-d308e78b7256-catalog-content\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.535904 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz6k2\" (UniqueName: \"kubernetes.io/projected/a2774890-f4e7-47b2-97c2-d308e78b7256-kube-api-access-tz6k2\") pod \"community-operators-xhnv8\" (UID: \"a2774890-f4e7-47b2-97c2-d308e78b7256\") " pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:53 crc kubenswrapper[4967]: I0120 08:55:53.670396 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.062204 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xhnv8"] Jan 20 08:55:54 crc kubenswrapper[4967]: W0120 08:55:54.119283 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2774890_f4e7_47b2_97c2_d308e78b7256.slice/crio-7b8df11fbb59201c5fd39751f1b78071711bfe97086a186142fd7045cc25ffc1 WatchSource:0}: Error finding container 7b8df11fbb59201c5fd39751f1b78071711bfe97086a186142fd7045cc25ffc1: Status 404 returned error can't find the container with id 7b8df11fbb59201c5fd39751f1b78071711bfe97086a186142fd7045cc25ffc1 Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.418999 4967 generic.go:334] "Generic (PLEG): container finished" podID="b1cfca35-a47d-4fc4-a425-57a60dd4b855" containerID="4d7ab0d1407a57bfeec88971a0d76644ac0ef184d1de9ec2d024e3c116233d67" exitCode=0 Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.419087 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q6d4w" event={"ID":"b1cfca35-a47d-4fc4-a425-57a60dd4b855","Type":"ContainerDied","Data":"4d7ab0d1407a57bfeec88971a0d76644ac0ef184d1de9ec2d024e3c116233d67"} Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.422238 4967 generic.go:334] "Generic (PLEG): container finished" podID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerID="86b8c57e65bbd3470c687010030f2558bdd380dfc15272e8859d48ace8f88a95" exitCode=0 Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.422310 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerDied","Data":"86b8c57e65bbd3470c687010030f2558bdd380dfc15272e8859d48ace8f88a95"} Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.422345 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerStarted","Data":"d614a956ee9c2f9f4a8c6e74a0731c8297c629f48081a445b0ba5112f3a50f94"} Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.427120 4967 generic.go:334] "Generic (PLEG): container finished" podID="a2774890-f4e7-47b2-97c2-d308e78b7256" containerID="562b0ed0135ec4138951044bf690ac31626de39465b4f01e553a5c643ef2174b" exitCode=0 Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.427163 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xhnv8" event={"ID":"a2774890-f4e7-47b2-97c2-d308e78b7256","Type":"ContainerDied","Data":"562b0ed0135ec4138951044bf690ac31626de39465b4f01e553a5c643ef2174b"} Jan 20 08:55:54 crc kubenswrapper[4967]: I0120 08:55:54.427199 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xhnv8" event={"ID":"a2774890-f4e7-47b2-97c2-d308e78b7256","Type":"ContainerStarted","Data":"7b8df11fbb59201c5fd39751f1b78071711bfe97086a186142fd7045cc25ffc1"} Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.120809 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f9v7m"] Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.122397 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.126224 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.136989 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9v7m"] Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.231746 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pckcc\" (UniqueName: \"kubernetes.io/projected/76857b90-e875-42ff-9332-a775e9cc5af6-kube-api-access-pckcc\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.231816 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-catalog-content\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.231839 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-utilities\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.333561 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pckcc\" (UniqueName: \"kubernetes.io/projected/76857b90-e875-42ff-9332-a775e9cc5af6-kube-api-access-pckcc\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.334263 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-catalog-content\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.334375 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-utilities\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.334848 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-catalog-content\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.335098 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76857b90-e875-42ff-9332-a775e9cc5af6-utilities\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.360911 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pckcc\" (UniqueName: \"kubernetes.io/projected/76857b90-e875-42ff-9332-a775e9cc5af6-kube-api-access-pckcc\") pod \"redhat-marketplace-f9v7m\" (UID: \"76857b90-e875-42ff-9332-a775e9cc5af6\") " pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.434182 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerStarted","Data":"160d3324cfae8fbba6d67a20085ad7ac24d9cf4944d6bf411a0d5ea7dfddc188"} Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.436911 4967 generic.go:334] "Generic (PLEG): container finished" podID="a2774890-f4e7-47b2-97c2-d308e78b7256" containerID="7b2fbf26024027eec259bea0a9c825ccd711d5c7fcfa42ce6e8820680cf0a038" exitCode=0 Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.436955 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xhnv8" event={"ID":"a2774890-f4e7-47b2-97c2-d308e78b7256","Type":"ContainerDied","Data":"7b2fbf26024027eec259bea0a9c825ccd711d5c7fcfa42ce6e8820680cf0a038"} Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.462733 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:55:55 crc kubenswrapper[4967]: I0120 08:55:55.874199 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9v7m"] Jan 20 08:55:55 crc kubenswrapper[4967]: W0120 08:55:55.883672 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76857b90_e875_42ff_9332_a775e9cc5af6.slice/crio-2bb76a501d5c670662f1085e0ac4bd4ffe0e997e44cbc002d84178c688221a4c WatchSource:0}: Error finding container 2bb76a501d5c670662f1085e0ac4bd4ffe0e997e44cbc002d84178c688221a4c: Status 404 returned error can't find the container with id 2bb76a501d5c670662f1085e0ac4bd4ffe0e997e44cbc002d84178c688221a4c Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.446416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q6d4w" event={"ID":"b1cfca35-a47d-4fc4-a425-57a60dd4b855","Type":"ContainerStarted","Data":"c735b737c51f49d1c2a2c3316dad3da8898157bbcf98f27174d19cdcace455d6"} Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.449661 4967 generic.go:334] "Generic (PLEG): container finished" podID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerID="160d3324cfae8fbba6d67a20085ad7ac24d9cf4944d6bf411a0d5ea7dfddc188" exitCode=0 Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.449724 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerDied","Data":"160d3324cfae8fbba6d67a20085ad7ac24d9cf4944d6bf411a0d5ea7dfddc188"} Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.454138 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xhnv8" event={"ID":"a2774890-f4e7-47b2-97c2-d308e78b7256","Type":"ContainerStarted","Data":"1a12b547852c2edeaff5b3060d5987c6f2c25d352baea2cc4ef67f0639796a2d"} Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.457003 4967 generic.go:334] "Generic (PLEG): container finished" podID="76857b90-e875-42ff-9332-a775e9cc5af6" containerID="c97ee166a2fd47c183caae4e4e7211bcafbb7135d8c45eb0d1f3f088b40457e9" exitCode=0 Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.457041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9v7m" event={"ID":"76857b90-e875-42ff-9332-a775e9cc5af6","Type":"ContainerDied","Data":"c97ee166a2fd47c183caae4e4e7211bcafbb7135d8c45eb0d1f3f088b40457e9"} Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.457065 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9v7m" event={"ID":"76857b90-e875-42ff-9332-a775e9cc5af6","Type":"ContainerStarted","Data":"2bb76a501d5c670662f1085e0ac4bd4ffe0e997e44cbc002d84178c688221a4c"} Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.501758 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q6d4w" podStartSLOduration=3.383869124 podStartE2EDuration="6.501742562s" podCreationTimestamp="2026-01-20 08:55:50 +0000 UTC" firstStartedPulling="2026-01-20 08:55:52.406412794 +0000 UTC m=+386.991713011" lastFinishedPulling="2026-01-20 08:55:55.524286242 +0000 UTC m=+390.109586449" observedRunningTime="2026-01-20 08:55:56.472559295 +0000 UTC m=+391.057859512" watchObservedRunningTime="2026-01-20 08:55:56.501742562 +0000 UTC m=+391.087042759" Jan 20 08:55:56 crc kubenswrapper[4967]: I0120 08:55:56.521486 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xhnv8" podStartSLOduration=1.996802018 podStartE2EDuration="3.521471431s" podCreationTimestamp="2026-01-20 08:55:53 +0000 UTC" firstStartedPulling="2026-01-20 08:55:54.430222166 +0000 UTC m=+389.015522373" lastFinishedPulling="2026-01-20 08:55:55.954891579 +0000 UTC m=+390.540191786" observedRunningTime="2026-01-20 08:55:56.519041649 +0000 UTC m=+391.104341856" watchObservedRunningTime="2026-01-20 08:55:56.521471431 +0000 UTC m=+391.106771638" Jan 20 08:55:57 crc kubenswrapper[4967]: I0120 08:55:57.463643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9v7m" event={"ID":"76857b90-e875-42ff-9332-a775e9cc5af6","Type":"ContainerStarted","Data":"1d75fdb7530474ff3c622fd68c97eb8200bf607edddfebfe88c9d565a5b6c2f6"} Jan 20 08:55:57 crc kubenswrapper[4967]: I0120 08:55:57.466496 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerStarted","Data":"8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2"} Jan 20 08:55:57 crc kubenswrapper[4967]: I0120 08:55:57.506345 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s5trl" podStartSLOduration=2.890612702 podStartE2EDuration="5.5063301s" podCreationTimestamp="2026-01-20 08:55:52 +0000 UTC" firstStartedPulling="2026-01-20 08:55:54.424709546 +0000 UTC m=+389.010009753" lastFinishedPulling="2026-01-20 08:55:57.040426944 +0000 UTC m=+391.625727151" observedRunningTime="2026-01-20 08:55:57.504268801 +0000 UTC m=+392.089568998" watchObservedRunningTime="2026-01-20 08:55:57.5063301 +0000 UTC m=+392.091630307" Jan 20 08:55:58 crc kubenswrapper[4967]: I0120 08:55:58.475576 4967 generic.go:334] "Generic (PLEG): container finished" podID="76857b90-e875-42ff-9332-a775e9cc5af6" containerID="1d75fdb7530474ff3c622fd68c97eb8200bf607edddfebfe88c9d565a5b6c2f6" exitCode=0 Jan 20 08:55:58 crc kubenswrapper[4967]: I0120 08:55:58.475642 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9v7m" event={"ID":"76857b90-e875-42ff-9332-a775e9cc5af6","Type":"ContainerDied","Data":"1d75fdb7530474ff3c622fd68c97eb8200bf607edddfebfe88c9d565a5b6c2f6"} Jan 20 08:56:00 crc kubenswrapper[4967]: I0120 08:56:00.490335 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9v7m" event={"ID":"76857b90-e875-42ff-9332-a775e9cc5af6","Type":"ContainerStarted","Data":"b8781030e11955d5b613ba6a0eeb7278a0ba0cb1b7ee9164dccfd54ad7fbffb3"} Jan 20 08:56:00 crc kubenswrapper[4967]: I0120 08:56:00.536398 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f9v7m" podStartSLOduration=1.9011368549999998 podStartE2EDuration="5.536375247s" podCreationTimestamp="2026-01-20 08:55:55 +0000 UTC" firstStartedPulling="2026-01-20 08:55:56.458339687 +0000 UTC m=+391.043639894" lastFinishedPulling="2026-01-20 08:56:00.093578039 +0000 UTC m=+394.678878286" observedRunningTime="2026-01-20 08:56:00.532454042 +0000 UTC m=+395.117754249" watchObservedRunningTime="2026-01-20 08:56:00.536375247 +0000 UTC m=+395.121675454" Jan 20 08:56:01 crc kubenswrapper[4967]: I0120 08:56:01.303942 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:56:01 crc kubenswrapper[4967]: I0120 08:56:01.304009 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:56:01 crc kubenswrapper[4967]: I0120 08:56:01.343719 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:56:01 crc kubenswrapper[4967]: I0120 08:56:01.536860 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q6d4w" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.043107 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.043702 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.094593 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.553334 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.671181 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.671242 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:56:03 crc kubenswrapper[4967]: I0120 08:56:03.725275 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:56:04 crc kubenswrapper[4967]: I0120 08:56:04.569269 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xhnv8" Jan 20 08:56:04 crc kubenswrapper[4967]: I0120 08:56:04.985475 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-d2hw7" Jan 20 08:56:05 crc kubenswrapper[4967]: I0120 08:56:05.053186 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:56:05 crc kubenswrapper[4967]: I0120 08:56:05.463648 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:56:05 crc kubenswrapper[4967]: I0120 08:56:05.463710 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:56:05 crc kubenswrapper[4967]: I0120 08:56:05.502113 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:56:05 crc kubenswrapper[4967]: I0120 08:56:05.563786 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f9v7m" Jan 20 08:56:18 crc kubenswrapper[4967]: I0120 08:56:18.475103 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:56:18 crc kubenswrapper[4967]: I0120 08:56:18.476152 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:56:18 crc kubenswrapper[4967]: I0120 08:56:18.476238 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:56:18 crc kubenswrapper[4967]: I0120 08:56:18.477226 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 08:56:18 crc kubenswrapper[4967]: I0120 08:56:18.477323 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6" gracePeriod=600 Jan 20 08:56:19 crc kubenswrapper[4967]: I0120 08:56:19.606565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6"} Jan 20 08:56:19 crc kubenswrapper[4967]: I0120 08:56:19.608839 4967 scope.go:117] "RemoveContainer" containerID="5140e5fa2dae3a15c67d0bc495f435756ef4e0429a65060a273990d28d3e510b" Jan 20 08:56:19 crc kubenswrapper[4967]: I0120 08:56:19.609309 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6" exitCode=0 Jan 20 08:56:19 crc kubenswrapper[4967]: I0120 08:56:19.609589 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd"} Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.087084 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" podUID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" containerName="registry" containerID="cri-o://f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3" gracePeriod=30 Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.529896 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654300 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654390 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654434 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654481 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654532 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvqw5\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654692 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654719 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.654769 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates\") pod \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\" (UID: \"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f\") " Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.655690 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.655817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.664062 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.664715 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.666167 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5" (OuterVolumeSpecName: "kube-api-access-jvqw5") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "kube-api-access-jvqw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.666311 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.666737 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.677367 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" (UID: "8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.687339 4967 generic.go:334] "Generic (PLEG): container finished" podID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" containerID="f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3" exitCode=0 Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.687395 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" event={"ID":"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f","Type":"ContainerDied","Data":"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3"} Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.687436 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.687466 4967 scope.go:117] "RemoveContainer" containerID="f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.687453 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8g5fl" event={"ID":"8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f","Type":"ContainerDied","Data":"fa2a077b630c362edff062ec1cbe9d5f8871077d0e06c0af1851f3c4e2265e34"} Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.723802 4967 scope.go:117] "RemoveContainer" containerID="f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3" Jan 20 08:56:30 crc kubenswrapper[4967]: E0120 08:56:30.724291 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3\": container with ID starting with f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3 not found: ID does not exist" containerID="f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.724347 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3"} err="failed to get container status \"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3\": rpc error: code = NotFound desc = could not find container \"f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3\": container with ID starting with f9385cad3773c62fc555ccc5db708d637743b691172ddbe3ccd914eaa99b73a3 not found: ID does not exist" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.739261 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.743483 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8g5fl"] Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756118 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756174 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756187 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvqw5\" (UniqueName: \"kubernetes.io/projected/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-kube-api-access-jvqw5\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756198 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756210 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756224 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:30 crc kubenswrapper[4967]: I0120 08:56:30.756237 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 20 08:56:31 crc kubenswrapper[4967]: I0120 08:56:31.701319 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" path="/var/lib/kubelet/pods/8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f/volumes" Jan 20 08:58:48 crc kubenswrapper[4967]: I0120 08:58:48.474551 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:58:48 crc kubenswrapper[4967]: I0120 08:58:48.475233 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:59:18 crc kubenswrapper[4967]: I0120 08:59:18.474562 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:59:18 crc kubenswrapper[4967]: I0120 08:59:18.475349 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:59:48 crc kubenswrapper[4967]: I0120 08:59:48.474491 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 08:59:48 crc kubenswrapper[4967]: I0120 08:59:48.475156 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 08:59:48 crc kubenswrapper[4967]: I0120 08:59:48.475219 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 08:59:48 crc kubenswrapper[4967]: I0120 08:59:48.476229 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 08:59:48 crc kubenswrapper[4967]: I0120 08:59:48.476329 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd" gracePeriod=600 Jan 20 08:59:49 crc kubenswrapper[4967]: I0120 08:59:49.001159 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd" exitCode=0 Jan 20 08:59:49 crc kubenswrapper[4967]: I0120 08:59:49.001267 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd"} Jan 20 08:59:49 crc kubenswrapper[4967]: I0120 08:59:49.001516 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840"} Jan 20 08:59:49 crc kubenswrapper[4967]: I0120 08:59:49.001533 4967 scope.go:117] "RemoveContainer" containerID="0385a32791206d4e1565f8779e8aab54e2f3d2e4677084a4632af3c1efd110b6" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.200793 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp"] Jan 20 09:00:00 crc kubenswrapper[4967]: E0120 09:00:00.201565 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" containerName="registry" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.201579 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" containerName="registry" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.201751 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8ac7e9-9ed9-412d-9dad-2a2c93be7b2f" containerName="registry" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.202238 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.204605 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.205983 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.207987 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp"] Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.376148 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w7rl\" (UniqueName: \"kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.376251 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.376285 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.476677 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w7rl\" (UniqueName: \"kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.476746 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.476765 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.477892 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.492992 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.495198 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w7rl\" (UniqueName: \"kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl\") pod \"collect-profiles-29481660-h8flp\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.524059 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:00 crc kubenswrapper[4967]: I0120 09:00:00.692455 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp"] Jan 20 09:00:01 crc kubenswrapper[4967]: I0120 09:00:01.078426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" event={"ID":"33decb60-8eac-41ba-bef7-a195406148bd","Type":"ContainerStarted","Data":"26a3ba83755f87585fe4374545087c0fefbec5148db5510c75468d0d66b696cd"} Jan 20 09:00:01 crc kubenswrapper[4967]: I0120 09:00:01.078488 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" event={"ID":"33decb60-8eac-41ba-bef7-a195406148bd","Type":"ContainerStarted","Data":"2e9309dce20606b9d4f2a53b5269489abfb8944f96f84520ab2c9dc1f296a835"} Jan 20 09:00:01 crc kubenswrapper[4967]: I0120 09:00:01.101550 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" podStartSLOduration=1.101530372 podStartE2EDuration="1.101530372s" podCreationTimestamp="2026-01-20 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:00:01.100924204 +0000 UTC m=+635.686224461" watchObservedRunningTime="2026-01-20 09:00:01.101530372 +0000 UTC m=+635.686830579" Jan 20 09:00:02 crc kubenswrapper[4967]: I0120 09:00:02.086240 4967 generic.go:334] "Generic (PLEG): container finished" podID="33decb60-8eac-41ba-bef7-a195406148bd" containerID="26a3ba83755f87585fe4374545087c0fefbec5148db5510c75468d0d66b696cd" exitCode=0 Jan 20 09:00:02 crc kubenswrapper[4967]: I0120 09:00:02.086305 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" event={"ID":"33decb60-8eac-41ba-bef7-a195406148bd","Type":"ContainerDied","Data":"26a3ba83755f87585fe4374545087c0fefbec5148db5510c75468d0d66b696cd"} Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.305751 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.411685 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume\") pod \"33decb60-8eac-41ba-bef7-a195406148bd\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.411808 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w7rl\" (UniqueName: \"kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl\") pod \"33decb60-8eac-41ba-bef7-a195406148bd\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.411877 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume\") pod \"33decb60-8eac-41ba-bef7-a195406148bd\" (UID: \"33decb60-8eac-41ba-bef7-a195406148bd\") " Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.412590 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "33decb60-8eac-41ba-bef7-a195406148bd" (UID: "33decb60-8eac-41ba-bef7-a195406148bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.430502 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "33decb60-8eac-41ba-bef7-a195406148bd" (UID: "33decb60-8eac-41ba-bef7-a195406148bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.430567 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl" (OuterVolumeSpecName: "kube-api-access-9w7rl") pod "33decb60-8eac-41ba-bef7-a195406148bd" (UID: "33decb60-8eac-41ba-bef7-a195406148bd"). InnerVolumeSpecName "kube-api-access-9w7rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.513008 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w7rl\" (UniqueName: \"kubernetes.io/projected/33decb60-8eac-41ba-bef7-a195406148bd-kube-api-access-9w7rl\") on node \"crc\" DevicePath \"\"" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.513048 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33decb60-8eac-41ba-bef7-a195406148bd-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:00:03 crc kubenswrapper[4967]: I0120 09:00:03.513059 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33decb60-8eac-41ba-bef7-a195406148bd-config-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:00:04 crc kubenswrapper[4967]: I0120 09:00:04.099285 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" event={"ID":"33decb60-8eac-41ba-bef7-a195406148bd","Type":"ContainerDied","Data":"2e9309dce20606b9d4f2a53b5269489abfb8944f96f84520ab2c9dc1f296a835"} Jan 20 09:00:04 crc kubenswrapper[4967]: I0120 09:00:04.099334 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e9309dce20606b9d4f2a53b5269489abfb8944f96f84520ab2c9dc1f296a835" Jan 20 09:00:04 crc kubenswrapper[4967]: I0120 09:00:04.099365 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481660-h8flp" Jan 20 09:01:48 crc kubenswrapper[4967]: I0120 09:01:48.475020 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:01:48 crc kubenswrapper[4967]: I0120 09:01:48.475754 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:02:01 crc kubenswrapper[4967]: I0120 09:02:01.555048 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 20 09:02:18 crc kubenswrapper[4967]: I0120 09:02:18.474669 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:02:18 crc kubenswrapper[4967]: I0120 09:02:18.475156 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:02:48 crc kubenswrapper[4967]: I0120 09:02:48.474327 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:02:48 crc kubenswrapper[4967]: I0120 09:02:48.474985 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:02:48 crc kubenswrapper[4967]: I0120 09:02:48.475045 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:02:48 crc kubenswrapper[4967]: I0120 09:02:48.475769 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:02:48 crc kubenswrapper[4967]: I0120 09:02:48.475835 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840" gracePeriod=600 Jan 20 09:02:49 crc kubenswrapper[4967]: I0120 09:02:49.044149 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840" exitCode=0 Jan 20 09:02:49 crc kubenswrapper[4967]: I0120 09:02:49.044235 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840"} Jan 20 09:02:49 crc kubenswrapper[4967]: I0120 09:02:49.044709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e"} Jan 20 09:02:49 crc kubenswrapper[4967]: I0120 09:02:49.044734 4967 scope.go:117] "RemoveContainer" containerID="f461e676a3d003c50a8ad160828a7b5f891d3099e92d04133c60c11d1a75e0bd" Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.459706 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc9bg"] Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.460841 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-controller" containerID="cri-o://6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461239 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="sbdb" containerID="cri-o://0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461285 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="nbdb" containerID="cri-o://081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461326 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="northd" containerID="cri-o://c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461366 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461410 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-node" containerID="cri-o://635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.461446 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-acl-logging" containerID="cri-o://8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4" gracePeriod=30 Jan 20 09:03:07 crc kubenswrapper[4967]: I0120 09:03:07.501257 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" containerID="cri-o://0ced0c42ffa0f0de4e4877f02fe69cf0b87ba9b8bd3cdfdd007227c3f57bad64" gracePeriod=30 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.150493 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/2.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.152287 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/1.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.152330 4967 generic.go:334] "Generic (PLEG): container finished" podID="7ba8630f-92bc-4708-a722-a7e27c747073" containerID="4238c896cd7352173c8ecf22ea04a431be302611922f554cfa608261d0b94127" exitCode=2 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.152405 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerDied","Data":"4238c896cd7352173c8ecf22ea04a431be302611922f554cfa608261d0b94127"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.152438 4967 scope.go:117] "RemoveContainer" containerID="124ba7a1e69b26048e04bf0327ca26991ea333331fc705f934fa75aae24aad41" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.152798 4967 scope.go:117] "RemoveContainer" containerID="4238c896cd7352173c8ecf22ea04a431be302611922f554cfa608261d0b94127" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.155150 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovnkube-controller/3.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.163376 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-acl-logging/0.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.164758 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-controller/0.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165264 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="0ced0c42ffa0f0de4e4877f02fe69cf0b87ba9b8bd3cdfdd007227c3f57bad64" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165294 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165305 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165316 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165324 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165332 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4" exitCode=0 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165340 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4" exitCode=143 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165348 4967 generic.go:334] "Generic (PLEG): container finished" podID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerID="6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5" exitCode=143 Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165343 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"0ced0c42ffa0f0de4e4877f02fe69cf0b87ba9b8bd3cdfdd007227c3f57bad64"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165400 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165420 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165432 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165444 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165457 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.165468 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5"} Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.186243 4967 scope.go:117] "RemoveContainer" containerID="51bd8c622474aa626dbc163a70b7b14cdcb12bc1b6360220775fc1fa849ecfeb" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.323098 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-acl-logging/0.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.323518 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-controller/0.log" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.323897 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354418 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354507 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354533 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354549 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354570 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354590 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash" (OuterVolumeSpecName: "host-slash") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354645 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354666 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354692 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354734 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354756 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354780 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354823 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354849 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354982 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355001 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355032 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354732 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354758 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354795 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log" (OuterVolumeSpecName: "node-log") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354823 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354849 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket" (OuterVolumeSpecName: "log-socket") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.354887 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355056 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355199 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49rck\" (UniqueName: \"kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355236 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355257 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") pod \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\" (UID: \"9e9f22bb-157c-4556-9066-cd3cdacdf3af\") " Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355644 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355663 4967 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-slash\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355673 4967 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355681 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355690 4967 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355700 4967 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355709 4967 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-log-socket\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355717 4967 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-node-log\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356100 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356204 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356280 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356313 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356408 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.355849 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356541 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356707 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.356761 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.362105 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.362793 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck" (OuterVolumeSpecName: "kube-api-access-49rck") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "kube-api-access-49rck". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.374283 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9e9f22bb-157c-4556-9066-cd3cdacdf3af" (UID: "9e9f22bb-157c-4556-9066-cd3cdacdf3af"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383124 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtwwl"] Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383342 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383355 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383363 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383369 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383382 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383387 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383395 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="northd" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383401 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="northd" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383408 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383415 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383423 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33decb60-8eac-41ba-bef7-a195406148bd" containerName="collect-profiles" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383428 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="33decb60-8eac-41ba-bef7-a195406148bd" containerName="collect-profiles" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383435 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-ovn-metrics" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383441 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-ovn-metrics" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383450 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-acl-logging" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383456 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-acl-logging" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383463 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-node" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383468 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-node" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383477 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="sbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383483 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="sbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383492 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="nbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383499 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="nbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383506 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383511 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383522 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kubecfg-setup" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383528 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kubecfg-setup" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383647 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383655 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="nbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383663 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383671 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-acl-logging" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383677 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovn-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383684 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-ovn-metrics" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383693 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383700 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="northd" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383709 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383717 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="33decb60-8eac-41ba-bef7-a195406148bd" containerName="collect-profiles" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383724 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383732 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="sbdb" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383740 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="kube-rbac-proxy-node" Jan 20 09:03:08 crc kubenswrapper[4967]: E0120 09:03:08.383817 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.383824 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" containerName="ovnkube-controller" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.385550 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457366 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-slash\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457429 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-netns\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457472 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-netd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457529 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-systemd-units\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457547 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-bin\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457589 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc6ln\" (UniqueName: \"kubernetes.io/projected/4d280a74-7f42-4ea4-b35b-4d8d042e9098-kube-api-access-jc6ln\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457671 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-log-socket\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457696 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-systemd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457721 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-kubelet\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457739 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-env-overrides\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457755 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-etc-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457788 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457805 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-ovn\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457842 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-var-lib-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457859 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovn-node-metrics-cert\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-script-lib\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457899 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-config\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457913 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-node-log\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457948 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457960 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457969 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457977 4967 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457986 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.457996 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49rck\" (UniqueName: \"kubernetes.io/projected/9e9f22bb-157c-4556-9066-cd3cdacdf3af-kube-api-access-49rck\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458005 4967 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458014 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458023 4967 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458031 4967 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458039 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f22bb-157c-4556-9066-cd3cdacdf3af-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.458047 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f22bb-157c-4556-9066-cd3cdacdf3af-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558777 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-slash\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558830 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-netns\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558865 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-netd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558889 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-systemd-units\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558911 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-bin\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558912 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-slash\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558934 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc6ln\" (UniqueName: \"kubernetes.io/projected/4d280a74-7f42-4ea4-b35b-4d8d042e9098-kube-api-access-jc6ln\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558981 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-netd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.558990 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-log-socket\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559047 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-log-socket\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559059 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-systemd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559091 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-netns\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559101 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-kubelet\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-systemd\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559129 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-env-overrides\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559151 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-kubelet\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559157 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559181 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-systemd-units\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559187 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559210 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-etc-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559233 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-ovn\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559255 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559285 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-var-lib-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559307 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-script-lib\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559325 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovn-node-metrics-cert\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559327 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-cni-bin\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-node-log\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559367 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-ovn\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559373 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-config\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559399 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559436 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.559437 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-etc-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560162 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-node-log\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560253 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-run-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560271 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d280a74-7f42-4ea4-b35b-4d8d042e9098-var-lib-openvswitch\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-env-overrides\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560995 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-config\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.560529 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovnkube-script-lib\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.563662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d280a74-7f42-4ea4-b35b-4d8d042e9098-ovn-node-metrics-cert\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.581703 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc6ln\" (UniqueName: \"kubernetes.io/projected/4d280a74-7f42-4ea4-b35b-4d8d042e9098-kube-api-access-jc6ln\") pod \"ovnkube-node-rtwwl\" (UID: \"4d280a74-7f42-4ea4-b35b-4d8d042e9098\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: I0120 09:03:08.700270 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:08 crc kubenswrapper[4967]: W0120 09:03:08.720808 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d280a74_7f42_4ea4_b35b_4d8d042e9098.slice/crio-28aab2a1ff5df9bed5588225f68f72287847aebfbd1d153f97d96397849b79cb WatchSource:0}: Error finding container 28aab2a1ff5df9bed5588225f68f72287847aebfbd1d153f97d96397849b79cb: Status 404 returned error can't find the container with id 28aab2a1ff5df9bed5588225f68f72287847aebfbd1d153f97d96397849b79cb Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.172720 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6wm4t_7ba8630f-92bc-4708-a722-a7e27c747073/kube-multus/2.log" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.173080 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6wm4t" event={"ID":"7ba8630f-92bc-4708-a722-a7e27c747073","Type":"ContainerStarted","Data":"78366e5e7709bb7ffb66560b004c7af13fc372633c810ed05c82857a6d9c18d0"} Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.182359 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-acl-logging/0.log" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.183031 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fc9bg_9e9f22bb-157c-4556-9066-cd3cdacdf3af/ovn-controller/0.log" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.183665 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" event={"ID":"9e9f22bb-157c-4556-9066-cd3cdacdf3af","Type":"ContainerDied","Data":"b32022fc51bdc7093b03b3236f1bb22dea985de51fe17320b6663cdcd3bf6564"} Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.183750 4967 scope.go:117] "RemoveContainer" containerID="0ced0c42ffa0f0de4e4877f02fe69cf0b87ba9b8bd3cdfdd007227c3f57bad64" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.183791 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fc9bg" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.185153 4967 generic.go:334] "Generic (PLEG): container finished" podID="4d280a74-7f42-4ea4-b35b-4d8d042e9098" containerID="65d18e6091cab00e41f1d0e6a7da2cbbc4bee41b7e695b9a33424c61c30243cb" exitCode=0 Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.185196 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerDied","Data":"65d18e6091cab00e41f1d0e6a7da2cbbc4bee41b7e695b9a33424c61c30243cb"} Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.185255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"28aab2a1ff5df9bed5588225f68f72287847aebfbd1d153f97d96397849b79cb"} Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.206225 4967 scope.go:117] "RemoveContainer" containerID="0a5609340627a96dead53fa877e61b9872c7cce170eea68fd9cc2e59a0e25d78" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.230881 4967 scope.go:117] "RemoveContainer" containerID="081aadd6eb939900153995d808c491ec871cefe538bb9c03682f7215cfeff61f" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.248929 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc9bg"] Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.253362 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fc9bg"] Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.265568 4967 scope.go:117] "RemoveContainer" containerID="c4ba7215d822a61ddb9e550539aaa0c7ec5e86fe307f17575105b3f7082587e7" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.284388 4967 scope.go:117] "RemoveContainer" containerID="c6a8d9220c30469656b08dcf337808545bb33657d2e4f3366c0b3662493a170e" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.297126 4967 scope.go:117] "RemoveContainer" containerID="635015290baf5307a8501ce969b47588ed8b77c98def8583bb97d766d52252a4" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.309494 4967 scope.go:117] "RemoveContainer" containerID="8d670e0494582df729fe747bd2a04263102b4ba1cee17a82ed3b405b27c554c4" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.324298 4967 scope.go:117] "RemoveContainer" containerID="6cdd71033ffe397b216b78961c231835fcd0b32c24fd4f8e162c5b5834978dd5" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.346996 4967 scope.go:117] "RemoveContainer" containerID="5a654f8da4cb967b11cee4db60c9661cc8afc0f55fc7cf2e2d8f3f3846bb0adb" Jan 20 09:03:09 crc kubenswrapper[4967]: I0120 09:03:09.700073 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9f22bb-157c-4556-9066-cd3cdacdf3af" path="/var/lib/kubelet/pods/9e9f22bb-157c-4556-9066-cd3cdacdf3af/volumes" Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.195763 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"8dd687d714990ceb2c9998801adb15dd454e3f3bd6445bf3dc8939c81a1d63cb"} Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.196035 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"c1fef2b5281d716bbbaab110b01037fb94d54e00d9945839fe07ecbd1a21e416"} Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.196053 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"a16526ca631a27c9e934697f0e1127f1b61a2f86a12a4befa37250ea6148a8ce"} Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.196066 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"d1d29cc028784c7a41201345d4a72138cd7ad79b841801a041ba6e8ed12ca72b"} Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.196077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"75ea89e7f49b74bf63f82805cafc7c83542c0ba2a7375cdef7acb1d1e2ebb799"} Jan 20 09:03:10 crc kubenswrapper[4967]: I0120 09:03:10.196089 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"45aef4fd155a79603fd127cb963f04e429b6417c15ed6e64701c2623ed3219fb"} Jan 20 09:03:13 crc kubenswrapper[4967]: I0120 09:03:13.217823 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"314b06e75c821b81d0c0ec884cde58bfe7d86ce076023b04592de60416bbce42"} Jan 20 09:03:16 crc kubenswrapper[4967]: I0120 09:03:16.238138 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" event={"ID":"4d280a74-7f42-4ea4-b35b-4d8d042e9098","Type":"ContainerStarted","Data":"225a5c32b6455b4d25c64c98b416bbeb4be6b690d2161a121d493bf9579982b6"} Jan 20 09:03:17 crc kubenswrapper[4967]: I0120 09:03:17.244143 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:17 crc kubenswrapper[4967]: I0120 09:03:17.244427 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:17 crc kubenswrapper[4967]: I0120 09:03:17.273319 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" podStartSLOduration=9.273302717 podStartE2EDuration="9.273302717s" podCreationTimestamp="2026-01-20 09:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:03:17.269795769 +0000 UTC m=+831.855096046" watchObservedRunningTime="2026-01-20 09:03:17.273302717 +0000 UTC m=+831.858602924" Jan 20 09:03:17 crc kubenswrapper[4967]: I0120 09:03:17.278126 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:18 crc kubenswrapper[4967]: I0120 09:03:18.250416 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:18 crc kubenswrapper[4967]: I0120 09:03:18.277113 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:38 crc kubenswrapper[4967]: I0120 09:03:38.722650 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtwwl" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.841275 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.843000 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.845512 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.846347 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-dpl6s" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.849214 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.858259 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.869173 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t4d5\" (UniqueName: \"kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5\") pod \"mariadb-operator-index-f7wc4\" (UID: \"82be7fc6-6c2c-4255-8da8-3874a329d22e\") " pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.970288 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t4d5\" (UniqueName: \"kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5\") pod \"mariadb-operator-index-f7wc4\" (UID: \"82be7fc6-6c2c-4255-8da8-3874a329d22e\") " pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:40 crc kubenswrapper[4967]: I0120 09:03:40.991366 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t4d5\" (UniqueName: \"kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5\") pod \"mariadb-operator-index-f7wc4\" (UID: \"82be7fc6-6c2c-4255-8da8-3874a329d22e\") " pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:41 crc kubenswrapper[4967]: I0120 09:03:41.160942 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:41 crc kubenswrapper[4967]: I0120 09:03:41.443450 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:41 crc kubenswrapper[4967]: I0120 09:03:41.449737 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 09:03:42 crc kubenswrapper[4967]: I0120 09:03:42.377044 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-f7wc4" event={"ID":"82be7fc6-6c2c-4255-8da8-3874a329d22e","Type":"ContainerStarted","Data":"ac23ea53ddb3a3ab69f9b9c27ec6737436ca956ecef30a1a211005148c99760f"} Jan 20 09:03:43 crc kubenswrapper[4967]: I0120 09:03:43.384915 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-f7wc4" event={"ID":"82be7fc6-6c2c-4255-8da8-3874a329d22e","Type":"ContainerStarted","Data":"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81"} Jan 20 09:03:43 crc kubenswrapper[4967]: I0120 09:03:43.818042 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-f7wc4" podStartSLOduration=3.021743936 podStartE2EDuration="3.818017905s" podCreationTimestamp="2026-01-20 09:03:40 +0000 UTC" firstStartedPulling="2026-01-20 09:03:41.449480165 +0000 UTC m=+856.034780372" lastFinishedPulling="2026-01-20 09:03:42.245754134 +0000 UTC m=+856.831054341" observedRunningTime="2026-01-20 09:03:43.403322827 +0000 UTC m=+857.988623074" watchObservedRunningTime="2026-01-20 09:03:43.818017905 +0000 UTC m=+858.403318112" Jan 20 09:03:43 crc kubenswrapper[4967]: I0120 09:03:43.818453 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.442675 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.443670 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.450645 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.509800 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8s2f\" (UniqueName: \"kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f\") pod \"mariadb-operator-index-lk86j\" (UID: \"45b78b67-e141-49df-ba8f-6e8f7958304b\") " pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.611654 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8s2f\" (UniqueName: \"kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f\") pod \"mariadb-operator-index-lk86j\" (UID: \"45b78b67-e141-49df-ba8f-6e8f7958304b\") " pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.632026 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8s2f\" (UniqueName: \"kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f\") pod \"mariadb-operator-index-lk86j\" (UID: \"45b78b67-e141-49df-ba8f-6e8f7958304b\") " pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.773796 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:44 crc kubenswrapper[4967]: I0120 09:03:44.969783 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.433288 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lk86j" event={"ID":"45b78b67-e141-49df-ba8f-6e8f7958304b","Type":"ContainerStarted","Data":"410d7d308a46be25b956ba97bc50717c51a78d7a265d9bf93160278d743f0d66"} Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.433392 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-f7wc4" podUID="82be7fc6-6c2c-4255-8da8-3874a329d22e" containerName="registry-server" containerID="cri-o://2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81" gracePeriod=2 Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.793700 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.826776 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t4d5\" (UniqueName: \"kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5\") pod \"82be7fc6-6c2c-4255-8da8-3874a329d22e\" (UID: \"82be7fc6-6c2c-4255-8da8-3874a329d22e\") " Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.832424 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5" (OuterVolumeSpecName: "kube-api-access-8t4d5") pod "82be7fc6-6c2c-4255-8da8-3874a329d22e" (UID: "82be7fc6-6c2c-4255-8da8-3874a329d22e"). InnerVolumeSpecName "kube-api-access-8t4d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:03:45 crc kubenswrapper[4967]: I0120 09:03:45.928480 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t4d5\" (UniqueName: \"kubernetes.io/projected/82be7fc6-6c2c-4255-8da8-3874a329d22e-kube-api-access-8t4d5\") on node \"crc\" DevicePath \"\"" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.442793 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lk86j" event={"ID":"45b78b67-e141-49df-ba8f-6e8f7958304b","Type":"ContainerStarted","Data":"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561"} Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.444792 4967 generic.go:334] "Generic (PLEG): container finished" podID="82be7fc6-6c2c-4255-8da8-3874a329d22e" containerID="2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81" exitCode=0 Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.444832 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-f7wc4" event={"ID":"82be7fc6-6c2c-4255-8da8-3874a329d22e","Type":"ContainerDied","Data":"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81"} Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.444835 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-f7wc4" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.444862 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-f7wc4" event={"ID":"82be7fc6-6c2c-4255-8da8-3874a329d22e","Type":"ContainerDied","Data":"ac23ea53ddb3a3ab69f9b9c27ec6737436ca956ecef30a1a211005148c99760f"} Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.444936 4967 scope.go:117] "RemoveContainer" containerID="2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.464570 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-lk86j" podStartSLOduration=1.872632774 podStartE2EDuration="2.464549448s" podCreationTimestamp="2026-01-20 09:03:44 +0000 UTC" firstStartedPulling="2026-01-20 09:03:44.980128214 +0000 UTC m=+859.565428431" lastFinishedPulling="2026-01-20 09:03:45.572044898 +0000 UTC m=+860.157345105" observedRunningTime="2026-01-20 09:03:46.461084721 +0000 UTC m=+861.046384968" watchObservedRunningTime="2026-01-20 09:03:46.464549448 +0000 UTC m=+861.049849655" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.464905 4967 scope.go:117] "RemoveContainer" containerID="2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81" Jan 20 09:03:46 crc kubenswrapper[4967]: E0120 09:03:46.465434 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81\": container with ID starting with 2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81 not found: ID does not exist" containerID="2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.465483 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81"} err="failed to get container status \"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81\": rpc error: code = NotFound desc = could not find container \"2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81\": container with ID starting with 2e93fd419bac34d418c254613daae9b2299ca4523ca5999ac6281fd6a7df5d81 not found: ID does not exist" Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.477905 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:46 crc kubenswrapper[4967]: I0120 09:03:46.480645 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-f7wc4"] Jan 20 09:03:47 crc kubenswrapper[4967]: I0120 09:03:47.702262 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82be7fc6-6c2c-4255-8da8-3874a329d22e" path="/var/lib/kubelet/pods/82be7fc6-6c2c-4255-8da8-3874a329d22e/volumes" Jan 20 09:03:54 crc kubenswrapper[4967]: I0120 09:03:54.774068 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:54 crc kubenswrapper[4967]: I0120 09:03:54.774457 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:54 crc kubenswrapper[4967]: I0120 09:03:54.814961 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:03:55 crc kubenswrapper[4967]: I0120 09:03:55.534810 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.664232 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l"] Jan 20 09:04:00 crc kubenswrapper[4967]: E0120 09:04:00.664758 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82be7fc6-6c2c-4255-8da8-3874a329d22e" containerName="registry-server" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.664773 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="82be7fc6-6c2c-4255-8da8-3874a329d22e" containerName="registry-server" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.664873 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="82be7fc6-6c2c-4255-8da8-3874a329d22e" containerName="registry-server" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.665577 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.671640 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lqxxp" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.675933 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l"] Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.732838 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph8h4\" (UniqueName: \"kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.732917 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.732943 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.833917 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph8h4\" (UniqueName: \"kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.833990 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.834023 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.834662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.834781 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.852979 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph8h4\" (UniqueName: \"kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4\") pod \"1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:00 crc kubenswrapper[4967]: I0120 09:04:00.983398 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:01 crc kubenswrapper[4967]: I0120 09:04:01.382047 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l"] Jan 20 09:04:01 crc kubenswrapper[4967]: W0120 09:04:01.389111 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod983b1332_9808_49d1_a25d_2a76ee5f9ce8.slice/crio-5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad WatchSource:0}: Error finding container 5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad: Status 404 returned error can't find the container with id 5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad Jan 20 09:04:01 crc kubenswrapper[4967]: I0120 09:04:01.550062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerStarted","Data":"1d0e51ee119e72be2320cbac89d3ddc1d45ee1d4d73e1caf6c7380f209db596e"} Jan 20 09:04:01 crc kubenswrapper[4967]: I0120 09:04:01.550406 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerStarted","Data":"5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad"} Jan 20 09:04:02 crc kubenswrapper[4967]: I0120 09:04:02.556006 4967 generic.go:334] "Generic (PLEG): container finished" podID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerID="1d0e51ee119e72be2320cbac89d3ddc1d45ee1d4d73e1caf6c7380f209db596e" exitCode=0 Jan 20 09:04:02 crc kubenswrapper[4967]: I0120 09:04:02.556088 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerDied","Data":"1d0e51ee119e72be2320cbac89d3ddc1d45ee1d4d73e1caf6c7380f209db596e"} Jan 20 09:04:04 crc kubenswrapper[4967]: I0120 09:04:04.568260 4967 generic.go:334] "Generic (PLEG): container finished" podID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerID="8781c51400de670a71aa7c6699e57908566242055de233b4237a6e9bafb676b4" exitCode=0 Jan 20 09:04:04 crc kubenswrapper[4967]: I0120 09:04:04.568353 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerDied","Data":"8781c51400de670a71aa7c6699e57908566242055de233b4237a6e9bafb676b4"} Jan 20 09:04:05 crc kubenswrapper[4967]: I0120 09:04:05.575930 4967 generic.go:334] "Generic (PLEG): container finished" podID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerID="793c635cc9edb7856b11ecec42c20a318a1c5ccb3167f75e629112d9bc413229" exitCode=0 Jan 20 09:04:05 crc kubenswrapper[4967]: I0120 09:04:05.575971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerDied","Data":"793c635cc9edb7856b11ecec42c20a318a1c5ccb3167f75e629112d9bc413229"} Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.796462 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.904758 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph8h4\" (UniqueName: \"kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4\") pod \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.905061 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util\") pod \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.905093 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle\") pod \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\" (UID: \"983b1332-9808-49d1-a25d-2a76ee5f9ce8\") " Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.906321 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle" (OuterVolumeSpecName: "bundle") pod "983b1332-9808-49d1-a25d-2a76ee5f9ce8" (UID: "983b1332-9808-49d1-a25d-2a76ee5f9ce8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.913012 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4" (OuterVolumeSpecName: "kube-api-access-ph8h4") pod "983b1332-9808-49d1-a25d-2a76ee5f9ce8" (UID: "983b1332-9808-49d1-a25d-2a76ee5f9ce8"). InnerVolumeSpecName "kube-api-access-ph8h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:04:06 crc kubenswrapper[4967]: I0120 09:04:06.927789 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util" (OuterVolumeSpecName: "util") pod "983b1332-9808-49d1-a25d-2a76ee5f9ce8" (UID: "983b1332-9808-49d1-a25d-2a76ee5f9ce8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.006799 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph8h4\" (UniqueName: \"kubernetes.io/projected/983b1332-9808-49d1-a25d-2a76ee5f9ce8-kube-api-access-ph8h4\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.006846 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-util\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.006860 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/983b1332-9808-49d1-a25d-2a76ee5f9ce8-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.591931 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" event={"ID":"983b1332-9808-49d1-a25d-2a76ee5f9ce8","Type":"ContainerDied","Data":"5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad"} Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.591981 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f0b7fa088c93b3a21d99546a2a6255d74b4973cbe1ddae3d82ac24f425f35ad" Jan 20 09:04:07 crc kubenswrapper[4967]: I0120 09:04:07.592005 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.350596 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:04:14 crc kubenswrapper[4967]: E0120 09:04:14.351042 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="util" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.351054 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="util" Jan 20 09:04:14 crc kubenswrapper[4967]: E0120 09:04:14.351063 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="pull" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.351072 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="pull" Jan 20 09:04:14 crc kubenswrapper[4967]: E0120 09:04:14.351081 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="extract" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.351086 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="extract" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.351174 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" containerName="extract" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.351518 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.354306 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-lr8vb" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.354557 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.355428 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.368267 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.394945 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.395006 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.395056 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrhnb\" (UniqueName: \"kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.496560 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrhnb\" (UniqueName: \"kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.496694 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.496726 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.502745 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.503158 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.511530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrhnb\" (UniqueName: \"kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb\") pod \"mariadb-operator-controller-manager-7f8fd5d49d-4bblk\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.667751 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:14 crc kubenswrapper[4967]: I0120 09:04:14.857464 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:04:15 crc kubenswrapper[4967]: I0120 09:04:15.629843 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerStarted","Data":"380a030fdf312547d4364c2ee0b46d539608cfbd8bb7044fcb9a2637718cdbc8"} Jan 20 09:04:21 crc kubenswrapper[4967]: I0120 09:04:21.662794 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerStarted","Data":"8946ebfdbb11e9d1b449ffdce1dcc75cdf64bec416e1d16252aaf05ccec4987f"} Jan 20 09:04:21 crc kubenswrapper[4967]: I0120 09:04:21.663352 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:21 crc kubenswrapper[4967]: I0120 09:04:21.698150 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" podStartSLOduration=1.652853677 podStartE2EDuration="7.698115701s" podCreationTimestamp="2026-01-20 09:04:14 +0000 UTC" firstStartedPulling="2026-01-20 09:04:14.868301126 +0000 UTC m=+889.453601333" lastFinishedPulling="2026-01-20 09:04:20.91356313 +0000 UTC m=+895.498863357" observedRunningTime="2026-01-20 09:04:21.692969197 +0000 UTC m=+896.278269454" watchObservedRunningTime="2026-01-20 09:04:21.698115701 +0000 UTC m=+896.283415948" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.101121 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.102975 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.121641 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.201507 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.201597 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.201673 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hhq\" (UniqueName: \"kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.302836 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.302906 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hhq\" (UniqueName: \"kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.302942 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.303394 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.303533 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.324303 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hhq\" (UniqueName: \"kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq\") pod \"community-operators-5gxnd\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.427897 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:28 crc kubenswrapper[4967]: I0120 09:04:28.701336 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:29 crc kubenswrapper[4967]: I0120 09:04:29.711993 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerDied","Data":"f7dbe39d280c30e985e98ed3ef8381af1e8c5a2102041dcc140d2a76a5a981be"} Jan 20 09:04:29 crc kubenswrapper[4967]: I0120 09:04:29.712345 4967 generic.go:334] "Generic (PLEG): container finished" podID="57495ace-1a95-4286-b715-63d8189f4c76" containerID="f7dbe39d280c30e985e98ed3ef8381af1e8c5a2102041dcc140d2a76a5a981be" exitCode=0 Jan 20 09:04:29 crc kubenswrapper[4967]: I0120 09:04:29.712367 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerStarted","Data":"b2c9223e4196a2b905e5fb9754db5241f632303b06b35a6d5947e29bcd5a39d1"} Jan 20 09:04:30 crc kubenswrapper[4967]: I0120 09:04:30.718259 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerStarted","Data":"23b4bf7166377f8bcdb4e4b65b0597605f2c1ac634d0fefe1ea6e7ee87bea9b5"} Jan 20 09:04:31 crc kubenswrapper[4967]: I0120 09:04:31.724939 4967 generic.go:334] "Generic (PLEG): container finished" podID="57495ace-1a95-4286-b715-63d8189f4c76" containerID="23b4bf7166377f8bcdb4e4b65b0597605f2c1ac634d0fefe1ea6e7ee87bea9b5" exitCode=0 Jan 20 09:04:31 crc kubenswrapper[4967]: I0120 09:04:31.725028 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerDied","Data":"23b4bf7166377f8bcdb4e4b65b0597605f2c1ac634d0fefe1ea6e7ee87bea9b5"} Jan 20 09:04:33 crc kubenswrapper[4967]: I0120 09:04:33.739488 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerStarted","Data":"1e70841b2f29a50046e2e36712c62f54d0cc421c33f9fe022b3128e839941ffc"} Jan 20 09:04:33 crc kubenswrapper[4967]: I0120 09:04:33.766406 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5gxnd" podStartSLOduration=2.220655209 podStartE2EDuration="5.76638951s" podCreationTimestamp="2026-01-20 09:04:28 +0000 UTC" firstStartedPulling="2026-01-20 09:04:29.715810856 +0000 UTC m=+904.301111103" lastFinishedPulling="2026-01-20 09:04:33.261545187 +0000 UTC m=+907.846845404" observedRunningTime="2026-01-20 09:04:33.762246622 +0000 UTC m=+908.347546839" watchObservedRunningTime="2026-01-20 09:04:33.76638951 +0000 UTC m=+908.351689717" Jan 20 09:04:34 crc kubenswrapper[4967]: I0120 09:04:34.672843 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:04:38 crc kubenswrapper[4967]: I0120 09:04:38.428168 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:38 crc kubenswrapper[4967]: I0120 09:04:38.428477 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:38 crc kubenswrapper[4967]: I0120 09:04:38.467486 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:38 crc kubenswrapper[4967]: I0120 09:04:38.799890 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:38 crc kubenswrapper[4967]: I0120 09:04:38.844501 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.341637 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c"] Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.342759 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.345479 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.355517 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c"] Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.459757 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5sgk\" (UniqueName: \"kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.460045 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.460092 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.561490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.561757 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5sgk\" (UniqueName: \"kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.561874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.562137 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.562287 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.581986 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5sgk\" (UniqueName: \"kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.660390 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.775711 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5gxnd" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="registry-server" containerID="cri-o://1e70841b2f29a50046e2e36712c62f54d0cc421c33f9fe022b3128e839941ffc" gracePeriod=2 Jan 20 09:04:40 crc kubenswrapper[4967]: I0120 09:04:40.861197 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c"] Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.768244 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerStarted","Data":"e0ad98f0e038c8b076d829e16be76142eaac1af77b48aa7cf82bcf748797e162"} Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.768649 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerStarted","Data":"03115c97839bac89a3ed4ee6e9708149c9fe7b2bf08e76eb7df92f65b6e8a8ac"} Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.772212 4967 generic.go:334] "Generic (PLEG): container finished" podID="57495ace-1a95-4286-b715-63d8189f4c76" containerID="1e70841b2f29a50046e2e36712c62f54d0cc421c33f9fe022b3128e839941ffc" exitCode=0 Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.772251 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerDied","Data":"1e70841b2f29a50046e2e36712c62f54d0cc421c33f9fe022b3128e839941ffc"} Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.792415 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.793792 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.814314 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.949480 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kftjd\" (UniqueName: \"kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.949576 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.949593 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:42 crc kubenswrapper[4967]: I0120 09:04:42.996470 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.051430 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kftjd\" (UniqueName: \"kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.051513 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.051532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.052008 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.052538 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.085737 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kftjd\" (UniqueName: \"kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd\") pod \"redhat-operators-5ps8k\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.133995 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.152221 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities\") pod \"57495ace-1a95-4286-b715-63d8189f4c76\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.152350 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4hhq\" (UniqueName: \"kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq\") pod \"57495ace-1a95-4286-b715-63d8189f4c76\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.152401 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content\") pod \"57495ace-1a95-4286-b715-63d8189f4c76\" (UID: \"57495ace-1a95-4286-b715-63d8189f4c76\") " Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.152965 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities" (OuterVolumeSpecName: "utilities") pod "57495ace-1a95-4286-b715-63d8189f4c76" (UID: "57495ace-1a95-4286-b715-63d8189f4c76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.158861 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq" (OuterVolumeSpecName: "kube-api-access-c4hhq") pod "57495ace-1a95-4286-b715-63d8189f4c76" (UID: "57495ace-1a95-4286-b715-63d8189f4c76"). InnerVolumeSpecName "kube-api-access-c4hhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.217819 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57495ace-1a95-4286-b715-63d8189f4c76" (UID: "57495ace-1a95-4286-b715-63d8189f4c76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.254073 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.254364 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57495ace-1a95-4286-b715-63d8189f4c76-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.254379 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4hhq\" (UniqueName: \"kubernetes.io/projected/57495ace-1a95-4286-b715-63d8189f4c76-kube-api-access-c4hhq\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.348590 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:04:43 crc kubenswrapper[4967]: W0120 09:04:43.357406 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf706bf15_8d26_40a5_8d70_e4a49399009f.slice/crio-e30512315e25b206093903a2c3a106e00f488368d0dda8fc16752cd925f30056 WatchSource:0}: Error finding container e30512315e25b206093903a2c3a106e00f488368d0dda8fc16752cd925f30056: Status 404 returned error can't find the container with id e30512315e25b206093903a2c3a106e00f488368d0dda8fc16752cd925f30056 Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.779270 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gxnd" event={"ID":"57495ace-1a95-4286-b715-63d8189f4c76","Type":"ContainerDied","Data":"b2c9223e4196a2b905e5fb9754db5241f632303b06b35a6d5947e29bcd5a39d1"} Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.780120 4967 scope.go:117] "RemoveContainer" containerID="1e70841b2f29a50046e2e36712c62f54d0cc421c33f9fe022b3128e839941ffc" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.779327 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gxnd" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.780698 4967 generic.go:334] "Generic (PLEG): container finished" podID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerID="528ef853a9631123099655efb0a70cee0819fc82230a06774f9b4756a73473ce" exitCode=0 Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.780748 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerDied","Data":"528ef853a9631123099655efb0a70cee0819fc82230a06774f9b4756a73473ce"} Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.780767 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerStarted","Data":"e30512315e25b206093903a2c3a106e00f488368d0dda8fc16752cd925f30056"} Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.786102 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de711c7-4972-4b89-8842-8d70a4126c13" containerID="e0ad98f0e038c8b076d829e16be76142eaac1af77b48aa7cf82bcf748797e162" exitCode=0 Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.786138 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerDied","Data":"e0ad98f0e038c8b076d829e16be76142eaac1af77b48aa7cf82bcf748797e162"} Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.795925 4967 scope.go:117] "RemoveContainer" containerID="23b4bf7166377f8bcdb4e4b65b0597605f2c1ac634d0fefe1ea6e7ee87bea9b5" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.814168 4967 scope.go:117] "RemoveContainer" containerID="f7dbe39d280c30e985e98ed3ef8381af1e8c5a2102041dcc140d2a76a5a981be" Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.837050 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:43 crc kubenswrapper[4967]: I0120 09:04:43.840667 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5gxnd"] Jan 20 09:04:45 crc kubenswrapper[4967]: I0120 09:04:45.700452 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57495ace-1a95-4286-b715-63d8189f4c76" path="/var/lib/kubelet/pods/57495ace-1a95-4286-b715-63d8189f4c76/volumes" Jan 20 09:04:45 crc kubenswrapper[4967]: I0120 09:04:45.802816 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de711c7-4972-4b89-8842-8d70a4126c13" containerID="80594a20a8892a167ac7b3b2a5031b8fbd7abaf23ba4f460b35a9bc69e3805e2" exitCode=0 Jan 20 09:04:45 crc kubenswrapper[4967]: I0120 09:04:45.802883 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerDied","Data":"80594a20a8892a167ac7b3b2a5031b8fbd7abaf23ba4f460b35a9bc69e3805e2"} Jan 20 09:04:45 crc kubenswrapper[4967]: I0120 09:04:45.806598 4967 generic.go:334] "Generic (PLEG): container finished" podID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerID="10bd51375d93abfc8bb7935304c73000826bf28218042504118f2a3540ce2b5b" exitCode=0 Jan 20 09:04:45 crc kubenswrapper[4967]: I0120 09:04:45.806647 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerDied","Data":"10bd51375d93abfc8bb7935304c73000826bf28218042504118f2a3540ce2b5b"} Jan 20 09:04:46 crc kubenswrapper[4967]: I0120 09:04:46.816098 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerDied","Data":"3ee66dc388b79f9d308e53a35e8cfbbfad3b4016272655936136abc02300131b"} Jan 20 09:04:46 crc kubenswrapper[4967]: I0120 09:04:46.816720 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de711c7-4972-4b89-8842-8d70a4126c13" containerID="3ee66dc388b79f9d308e53a35e8cfbbfad3b4016272655936136abc02300131b" exitCode=0 Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.082569 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.215972 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle\") pod \"5de711c7-4972-4b89-8842-8d70a4126c13\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.216045 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5sgk\" (UniqueName: \"kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk\") pod \"5de711c7-4972-4b89-8842-8d70a4126c13\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.216082 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util\") pod \"5de711c7-4972-4b89-8842-8d70a4126c13\" (UID: \"5de711c7-4972-4b89-8842-8d70a4126c13\") " Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.218340 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle" (OuterVolumeSpecName: "bundle") pod "5de711c7-4972-4b89-8842-8d70a4126c13" (UID: "5de711c7-4972-4b89-8842-8d70a4126c13"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.221873 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk" (OuterVolumeSpecName: "kube-api-access-v5sgk") pod "5de711c7-4972-4b89-8842-8d70a4126c13" (UID: "5de711c7-4972-4b89-8842-8d70a4126c13"). InnerVolumeSpecName "kube-api-access-v5sgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.226379 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util" (OuterVolumeSpecName: "util") pod "5de711c7-4972-4b89-8842-8d70a4126c13" (UID: "5de711c7-4972-4b89-8842-8d70a4126c13"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.317554 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.317590 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5sgk\" (UniqueName: \"kubernetes.io/projected/5de711c7-4972-4b89-8842-8d70a4126c13-kube-api-access-v5sgk\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.317600 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5de711c7-4972-4b89-8842-8d70a4126c13-util\") on node \"crc\" DevicePath \"\"" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.474441 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.474498 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.828988 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerStarted","Data":"fe36b0bfde7999bab0b0bea24ea18481a53c0dde5824d4e3ebe724d3eb8316df"} Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.831216 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" event={"ID":"5de711c7-4972-4b89-8842-8d70a4126c13","Type":"ContainerDied","Data":"03115c97839bac89a3ed4ee6e9708149c9fe7b2bf08e76eb7df92f65b6e8a8ac"} Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.831261 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03115c97839bac89a3ed4ee6e9708149c9fe7b2bf08e76eb7df92f65b6e8a8ac" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.831351 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c" Jan 20 09:04:48 crc kubenswrapper[4967]: I0120 09:04:48.849580 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5ps8k" podStartSLOduration=2.8203324800000003 podStartE2EDuration="6.849560672s" podCreationTimestamp="2026-01-20 09:04:42 +0000 UTC" firstStartedPulling="2026-01-20 09:04:43.782039085 +0000 UTC m=+918.367339292" lastFinishedPulling="2026-01-20 09:04:47.811267277 +0000 UTC m=+922.396567484" observedRunningTime="2026-01-20 09:04:48.847833473 +0000 UTC m=+923.433133680" watchObservedRunningTime="2026-01-20 09:04:48.849560672 +0000 UTC m=+923.434860879" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.134364 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.134710 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.706858 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707433 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="pull" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707450 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="pull" Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707463 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="registry-server" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707471 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="registry-server" Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707483 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="util" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707490 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="util" Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707502 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="extract" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707509 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="extract" Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707523 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="extract-utilities" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707530 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="extract-utilities" Jan 20 09:04:53 crc kubenswrapper[4967]: E0120 09:04:53.707542 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="extract-content" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707553 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="extract-content" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707695 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de711c7-4972-4b89-8842-8d70a4126c13" containerName="extract" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.707710 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="57495ace-1a95-4286-b715-63d8189f4c76" containerName="registry-server" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.708706 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.717201 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.885585 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.885664 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.885688 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nd4c\" (UniqueName: \"kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.986889 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.986929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nd4c\" (UniqueName: \"kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.987000 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.987434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:53 crc kubenswrapper[4967]: I0120 09:04:53.987470 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:54 crc kubenswrapper[4967]: I0120 09:04:54.015036 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nd4c\" (UniqueName: \"kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c\") pod \"certified-operators-bgxhj\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:54 crc kubenswrapper[4967]: I0120 09:04:54.024551 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:04:54 crc kubenswrapper[4967]: I0120 09:04:54.179487 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5ps8k" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="registry-server" probeResult="failure" output=< Jan 20 09:04:54 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 09:04:54 crc kubenswrapper[4967]: > Jan 20 09:04:54 crc kubenswrapper[4967]: I0120 09:04:54.211415 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:04:54 crc kubenswrapper[4967]: I0120 09:04:54.868607 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerStarted","Data":"08ec42bf9295d8d57832c549da035ba9913775d0c395bb10dedfdd477a30c7eb"} Jan 20 09:04:56 crc kubenswrapper[4967]: I0120 09:04:56.879379 4967 generic.go:334] "Generic (PLEG): container finished" podID="156364ad-181d-441d-9754-104d3a686a3a" containerID="8be5e04ce4ecd67bc011e15f6503b6fbde15ff35197970d2fbf4249b1cefa217" exitCode=0 Jan 20 09:04:56 crc kubenswrapper[4967]: I0120 09:04:56.879529 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerDied","Data":"8be5e04ce4ecd67bc011e15f6503b6fbde15ff35197970d2fbf4249b1cefa217"} Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.152166 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc"] Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.152858 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.157818 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.158149 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.158796 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rgrbz" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.158943 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.160491 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.174550 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc"] Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.333426 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-apiservice-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.333517 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk42l\" (UniqueName: \"kubernetes.io/projected/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-kube-api-access-nk42l\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.333543 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-webhook-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.434910 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-apiservice-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.434974 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk42l\" (UniqueName: \"kubernetes.io/projected/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-kube-api-access-nk42l\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.434999 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-webhook-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.449067 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-webhook-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.450358 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-apiservice-cert\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.466421 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk42l\" (UniqueName: \"kubernetes.io/projected/ccc3a123-c88b-42ae-8c10-5fc70762b0a3-kube-api-access-nk42l\") pod \"metallb-operator-controller-manager-9c99dbd64-q2zrc\" (UID: \"ccc3a123-c88b-42ae-8c10-5fc70762b0a3\") " pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.471424 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.507343 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv"] Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.508713 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.510388 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.510526 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-z7xwt" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.511198 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.623121 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv"] Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.637734 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-apiservice-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.637791 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l4r6\" (UniqueName: \"kubernetes.io/projected/96b5beda-b9ed-4f48-9dcb-8a261f532b47-kube-api-access-7l4r6\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.637817 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-webhook-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.738561 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-webhook-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.738722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-apiservice-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.738765 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l4r6\" (UniqueName: \"kubernetes.io/projected/96b5beda-b9ed-4f48-9dcb-8a261f532b47-kube-api-access-7l4r6\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.744290 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-webhook-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.750938 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96b5beda-b9ed-4f48-9dcb-8a261f532b47-apiservice-cert\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.754882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l4r6\" (UniqueName: \"kubernetes.io/projected/96b5beda-b9ed-4f48-9dcb-8a261f532b47-kube-api-access-7l4r6\") pod \"metallb-operator-webhook-server-74dcbc6877-8v4vv\" (UID: \"96b5beda-b9ed-4f48-9dcb-8a261f532b47\") " pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:57 crc kubenswrapper[4967]: I0120 09:04:57.973962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:04:58 crc kubenswrapper[4967]: I0120 09:04:58.036187 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc"] Jan 20 09:04:58 crc kubenswrapper[4967]: W0120 09:04:58.043269 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccc3a123_c88b_42ae_8c10_5fc70762b0a3.slice/crio-bac5fc45b01b4ee1ae0a77914649747f183faa67f6c0f1e3d9de40b57a82dd95 WatchSource:0}: Error finding container bac5fc45b01b4ee1ae0a77914649747f183faa67f6c0f1e3d9de40b57a82dd95: Status 404 returned error can't find the container with id bac5fc45b01b4ee1ae0a77914649747f183faa67f6c0f1e3d9de40b57a82dd95 Jan 20 09:04:58 crc kubenswrapper[4967]: I0120 09:04:58.178676 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv"] Jan 20 09:04:58 crc kubenswrapper[4967]: I0120 09:04:58.892469 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" event={"ID":"ccc3a123-c88b-42ae-8c10-5fc70762b0a3","Type":"ContainerStarted","Data":"bac5fc45b01b4ee1ae0a77914649747f183faa67f6c0f1e3d9de40b57a82dd95"} Jan 20 09:04:58 crc kubenswrapper[4967]: I0120 09:04:58.893467 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" event={"ID":"96b5beda-b9ed-4f48-9dcb-8a261f532b47","Type":"ContainerStarted","Data":"cf916f3bcabfb6f54d9f7e5561dad3f1444284a4b53789ce2cafa91495bda2fb"} Jan 20 09:05:01 crc kubenswrapper[4967]: I0120 09:05:01.911165 4967 generic.go:334] "Generic (PLEG): container finished" podID="156364ad-181d-441d-9754-104d3a686a3a" containerID="df1aa7045322b9ba584911b4b2d8024cbe93aa4f5e741d37092abcb29bcdc51b" exitCode=0 Jan 20 09:05:01 crc kubenswrapper[4967]: I0120 09:05:01.911236 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerDied","Data":"df1aa7045322b9ba584911b4b2d8024cbe93aa4f5e741d37092abcb29bcdc51b"} Jan 20 09:05:03 crc kubenswrapper[4967]: I0120 09:05:03.188813 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:05:03 crc kubenswrapper[4967]: I0120 09:05:03.230074 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:05:06 crc kubenswrapper[4967]: I0120 09:05:06.698660 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:05:06 crc kubenswrapper[4967]: I0120 09:05:06.699430 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5ps8k" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="registry-server" containerID="cri-o://fe36b0bfde7999bab0b0bea24ea18481a53c0dde5824d4e3ebe724d3eb8316df" gracePeriod=2 Jan 20 09:05:06 crc kubenswrapper[4967]: I0120 09:05:06.944096 4967 generic.go:334] "Generic (PLEG): container finished" podID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerID="fe36b0bfde7999bab0b0bea24ea18481a53c0dde5824d4e3ebe724d3eb8316df" exitCode=0 Jan 20 09:05:06 crc kubenswrapper[4967]: I0120 09:05:06.944137 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerDied","Data":"fe36b0bfde7999bab0b0bea24ea18481a53c0dde5824d4e3ebe724d3eb8316df"} Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.711271 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.712796 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.728765 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.791767 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.791898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.791951 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.893439 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.893532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.893589 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.894370 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.894675 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:07 crc kubenswrapper[4967]: I0120 09:05:07.919510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp\") pod \"redhat-marketplace-x8nzf\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:08 crc kubenswrapper[4967]: I0120 09:05:08.035787 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:09 crc kubenswrapper[4967]: I0120 09:05:09.913809 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:05:09 crc kubenswrapper[4967]: I0120 09:05:09.984827 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ps8k" event={"ID":"f706bf15-8d26-40a5-8d70-e4a49399009f","Type":"ContainerDied","Data":"e30512315e25b206093903a2c3a106e00f488368d0dda8fc16752cd925f30056"} Jan 20 09:05:09 crc kubenswrapper[4967]: I0120 09:05:09.984907 4967 scope.go:117] "RemoveContainer" containerID="fe36b0bfde7999bab0b0bea24ea18481a53c0dde5824d4e3ebe724d3eb8316df" Jan 20 09:05:09 crc kubenswrapper[4967]: I0120 09:05:09.985088 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ps8k" Jan 20 09:05:09 crc kubenswrapper[4967]: I0120 09:05:09.996207 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:10 crc kubenswrapper[4967]: W0120 09:05:10.003138 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dce87bd_d333_4e97_b31c_c003889108a6.slice/crio-f841dfadfb8bb73b0c7c9615bb31d9ec5c9ea2187994efbefb5ec884c931c32d WatchSource:0}: Error finding container f841dfadfb8bb73b0c7c9615bb31d9ec5c9ea2187994efbefb5ec884c931c32d: Status 404 returned error can't find the container with id f841dfadfb8bb73b0c7c9615bb31d9ec5c9ea2187994efbefb5ec884c931c32d Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.008379 4967 scope.go:117] "RemoveContainer" containerID="10bd51375d93abfc8bb7935304c73000826bf28218042504118f2a3540ce2b5b" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.025357 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kftjd\" (UniqueName: \"kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd\") pod \"f706bf15-8d26-40a5-8d70-e4a49399009f\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.025451 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content\") pod \"f706bf15-8d26-40a5-8d70-e4a49399009f\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.025518 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities\") pod \"f706bf15-8d26-40a5-8d70-e4a49399009f\" (UID: \"f706bf15-8d26-40a5-8d70-e4a49399009f\") " Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.026476 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities" (OuterVolumeSpecName: "utilities") pod "f706bf15-8d26-40a5-8d70-e4a49399009f" (UID: "f706bf15-8d26-40a5-8d70-e4a49399009f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.030708 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd" (OuterVolumeSpecName: "kube-api-access-kftjd") pod "f706bf15-8d26-40a5-8d70-e4a49399009f" (UID: "f706bf15-8d26-40a5-8d70-e4a49399009f"). InnerVolumeSpecName "kube-api-access-kftjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.036364 4967 scope.go:117] "RemoveContainer" containerID="528ef853a9631123099655efb0a70cee0819fc82230a06774f9b4756a73473ce" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.126957 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kftjd\" (UniqueName: \"kubernetes.io/projected/f706bf15-8d26-40a5-8d70-e4a49399009f-kube-api-access-kftjd\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.126985 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.161208 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f706bf15-8d26-40a5-8d70-e4a49399009f" (UID: "f706bf15-8d26-40a5-8d70-e4a49399009f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.227945 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f706bf15-8d26-40a5-8d70-e4a49399009f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.311888 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.318014 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5ps8k"] Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.991935 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerStarted","Data":"ff235afc6e76d26e89a3c0d4c702e83846e449ba420fabc16ce9fb46f894e577"} Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.993461 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" event={"ID":"ccc3a123-c88b-42ae-8c10-5fc70762b0a3","Type":"ContainerStarted","Data":"ba252b4e5e651d7389b6e3bd6e4bb457c79093e104013896d4b5eb66959486b9"} Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.993601 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.995862 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" event={"ID":"96b5beda-b9ed-4f48-9dcb-8a261f532b47","Type":"ContainerStarted","Data":"33bc70a7b4ddb6c5f9a8609cb2032b5116a6be3dcca376f750a343a634b434a6"} Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.996459 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.997596 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dce87bd-d333-4e97-b31c-c003889108a6" containerID="d8e6c57a021aea3472128dbc126c83c00425480d37cf8f3d1e315e64a0d57fe8" exitCode=0 Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.997646 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerDied","Data":"d8e6c57a021aea3472128dbc126c83c00425480d37cf8f3d1e315e64a0d57fe8"} Jan 20 09:05:10 crc kubenswrapper[4967]: I0120 09:05:10.997662 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerStarted","Data":"f841dfadfb8bb73b0c7c9615bb31d9ec5c9ea2187994efbefb5ec884c931c32d"} Jan 20 09:05:11 crc kubenswrapper[4967]: I0120 09:05:11.019909 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bgxhj" podStartSLOduration=5.058303127 podStartE2EDuration="18.019890302s" podCreationTimestamp="2026-01-20 09:04:53 +0000 UTC" firstStartedPulling="2026-01-20 09:04:56.880771147 +0000 UTC m=+931.466071354" lastFinishedPulling="2026-01-20 09:05:09.842358312 +0000 UTC m=+944.427658529" observedRunningTime="2026-01-20 09:05:11.018928075 +0000 UTC m=+945.604228292" watchObservedRunningTime="2026-01-20 09:05:11.019890302 +0000 UTC m=+945.605190509" Jan 20 09:05:11 crc kubenswrapper[4967]: I0120 09:05:11.038651 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" podStartSLOduration=2.332744499 podStartE2EDuration="14.038633231s" podCreationTimestamp="2026-01-20 09:04:57 +0000 UTC" firstStartedPulling="2026-01-20 09:04:58.046955027 +0000 UTC m=+932.632255234" lastFinishedPulling="2026-01-20 09:05:09.752843749 +0000 UTC m=+944.338143966" observedRunningTime="2026-01-20 09:05:11.038454836 +0000 UTC m=+945.623755043" watchObservedRunningTime="2026-01-20 09:05:11.038633231 +0000 UTC m=+945.623933438" Jan 20 09:05:11 crc kubenswrapper[4967]: I0120 09:05:11.058316 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" podStartSLOduration=2.4905703790000002 podStartE2EDuration="14.058291985s" podCreationTimestamp="2026-01-20 09:04:57 +0000 UTC" firstStartedPulling="2026-01-20 09:04:58.186853621 +0000 UTC m=+932.772153828" lastFinishedPulling="2026-01-20 09:05:09.754575227 +0000 UTC m=+944.339875434" observedRunningTime="2026-01-20 09:05:11.053230902 +0000 UTC m=+945.638531109" watchObservedRunningTime="2026-01-20 09:05:11.058291985 +0000 UTC m=+945.643592192" Jan 20 09:05:11 crc kubenswrapper[4967]: I0120 09:05:11.700089 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" path="/var/lib/kubelet/pods/f706bf15-8d26-40a5-8d70-e4a49399009f/volumes" Jan 20 09:05:13 crc kubenswrapper[4967]: I0120 09:05:13.010868 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dce87bd-d333-4e97-b31c-c003889108a6" containerID="23effd996a8d43d3cd1fc839dc0ab4847e088edd94fc1f4c43f901e3fb177ebf" exitCode=0 Jan 20 09:05:13 crc kubenswrapper[4967]: I0120 09:05:13.010962 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerDied","Data":"23effd996a8d43d3cd1fc839dc0ab4847e088edd94fc1f4c43f901e3fb177ebf"} Jan 20 09:05:14 crc kubenswrapper[4967]: I0120 09:05:14.024840 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:14 crc kubenswrapper[4967]: I0120 09:05:14.024897 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:14 crc kubenswrapper[4967]: I0120 09:05:14.062983 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:15 crc kubenswrapper[4967]: I0120 09:05:15.025858 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerStarted","Data":"df97d9cfeb876e583bb827fa98d6a66676a942e425241baa3caceb849cac3927"} Jan 20 09:05:15 crc kubenswrapper[4967]: I0120 09:05:15.047930 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x8nzf" podStartSLOduration=4.538807444 podStartE2EDuration="8.04790721s" podCreationTimestamp="2026-01-20 09:05:07 +0000 UTC" firstStartedPulling="2026-01-20 09:05:10.99923215 +0000 UTC m=+945.584532357" lastFinishedPulling="2026-01-20 09:05:14.508331916 +0000 UTC m=+949.093632123" observedRunningTime="2026-01-20 09:05:15.043885566 +0000 UTC m=+949.629185783" watchObservedRunningTime="2026-01-20 09:05:15.04790721 +0000 UTC m=+949.633207417" Jan 20 09:05:15 crc kubenswrapper[4967]: I0120 09:05:15.066478 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:16 crc kubenswrapper[4967]: I0120 09:05:16.096448 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:05:17 crc kubenswrapper[4967]: I0120 09:05:17.035417 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bgxhj" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="registry-server" containerID="cri-o://ff235afc6e76d26e89a3c0d4c702e83846e449ba420fabc16ce9fb46f894e577" gracePeriod=2 Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.037000 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.037369 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.042720 4967 generic.go:334] "Generic (PLEG): container finished" podID="156364ad-181d-441d-9754-104d3a686a3a" containerID="ff235afc6e76d26e89a3c0d4c702e83846e449ba420fabc16ce9fb46f894e577" exitCode=0 Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.042780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerDied","Data":"ff235afc6e76d26e89a3c0d4c702e83846e449ba420fabc16ce9fb46f894e577"} Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.077508 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.112233 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.240397 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content\") pod \"156364ad-181d-441d-9754-104d3a686a3a\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.240450 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nd4c\" (UniqueName: \"kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c\") pod \"156364ad-181d-441d-9754-104d3a686a3a\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.240515 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities\") pod \"156364ad-181d-441d-9754-104d3a686a3a\" (UID: \"156364ad-181d-441d-9754-104d3a686a3a\") " Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.241464 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities" (OuterVolumeSpecName: "utilities") pod "156364ad-181d-441d-9754-104d3a686a3a" (UID: "156364ad-181d-441d-9754-104d3a686a3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.245926 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c" (OuterVolumeSpecName: "kube-api-access-2nd4c") pod "156364ad-181d-441d-9754-104d3a686a3a" (UID: "156364ad-181d-441d-9754-104d3a686a3a"). InnerVolumeSpecName "kube-api-access-2nd4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.285134 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "156364ad-181d-441d-9754-104d3a686a3a" (UID: "156364ad-181d-441d-9754-104d3a686a3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.341889 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.341932 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nd4c\" (UniqueName: \"kubernetes.io/projected/156364ad-181d-441d-9754-104d3a686a3a-kube-api-access-2nd4c\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.341949 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156364ad-181d-441d-9754-104d3a686a3a-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.474726 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:05:18 crc kubenswrapper[4967]: I0120 09:05:18.474784 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.050197 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bgxhj" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.050195 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bgxhj" event={"ID":"156364ad-181d-441d-9754-104d3a686a3a","Type":"ContainerDied","Data":"08ec42bf9295d8d57832c549da035ba9913775d0c395bb10dedfdd477a30c7eb"} Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.050256 4967 scope.go:117] "RemoveContainer" containerID="ff235afc6e76d26e89a3c0d4c702e83846e449ba420fabc16ce9fb46f894e577" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.064534 4967 scope.go:117] "RemoveContainer" containerID="df1aa7045322b9ba584911b4b2d8024cbe93aa4f5e741d37092abcb29bcdc51b" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.080518 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.087576 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bgxhj"] Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.096121 4967 scope.go:117] "RemoveContainer" containerID="8be5e04ce4ecd67bc011e15f6503b6fbde15ff35197970d2fbf4249b1cefa217" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.113737 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:19 crc kubenswrapper[4967]: I0120 09:05:19.700672 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156364ad-181d-441d-9754-104d3a686a3a" path="/var/lib/kubelet/pods/156364ad-181d-441d-9754-104d3a686a3a/volumes" Jan 20 09:05:23 crc kubenswrapper[4967]: I0120 09:05:23.097131 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:23 crc kubenswrapper[4967]: I0120 09:05:23.097356 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x8nzf" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="registry-server" containerID="cri-o://df97d9cfeb876e583bb827fa98d6a66676a942e425241baa3caceb849cac3927" gracePeriod=2 Jan 20 09:05:25 crc kubenswrapper[4967]: I0120 09:05:25.088443 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dce87bd-d333-4e97-b31c-c003889108a6" containerID="df97d9cfeb876e583bb827fa98d6a66676a942e425241baa3caceb849cac3927" exitCode=0 Jan 20 09:05:25 crc kubenswrapper[4967]: I0120 09:05:25.088725 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerDied","Data":"df97d9cfeb876e583bb827fa98d6a66676a942e425241baa3caceb849cac3927"} Jan 20 09:05:25 crc kubenswrapper[4967]: I0120 09:05:25.991395 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.096875 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8nzf" event={"ID":"3dce87bd-d333-4e97-b31c-c003889108a6","Type":"ContainerDied","Data":"f841dfadfb8bb73b0c7c9615bb31d9ec5c9ea2187994efbefb5ec884c931c32d"} Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.096946 4967 scope.go:117] "RemoveContainer" containerID="df97d9cfeb876e583bb827fa98d6a66676a942e425241baa3caceb849cac3927" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.096963 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8nzf" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.117538 4967 scope.go:117] "RemoveContainer" containerID="23effd996a8d43d3cd1fc839dc0ab4847e088edd94fc1f4c43f901e3fb177ebf" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.130941 4967 scope.go:117] "RemoveContainer" containerID="d8e6c57a021aea3472128dbc126c83c00425480d37cf8f3d1e315e64a0d57fe8" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.135684 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp\") pod \"3dce87bd-d333-4e97-b31c-c003889108a6\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.135799 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities\") pod \"3dce87bd-d333-4e97-b31c-c003889108a6\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.135836 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content\") pod \"3dce87bd-d333-4e97-b31c-c003889108a6\" (UID: \"3dce87bd-d333-4e97-b31c-c003889108a6\") " Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.137358 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities" (OuterVolumeSpecName: "utilities") pod "3dce87bd-d333-4e97-b31c-c003889108a6" (UID: "3dce87bd-d333-4e97-b31c-c003889108a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.145085 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp" (OuterVolumeSpecName: "kube-api-access-9ltzp") pod "3dce87bd-d333-4e97-b31c-c003889108a6" (UID: "3dce87bd-d333-4e97-b31c-c003889108a6"). InnerVolumeSpecName "kube-api-access-9ltzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.168193 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dce87bd-d333-4e97-b31c-c003889108a6" (UID: "3dce87bd-d333-4e97-b31c-c003889108a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.236995 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.237038 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dce87bd-d333-4e97-b31c-c003889108a6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.237054 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/3dce87bd-d333-4e97-b31c-c003889108a6-kube-api-access-9ltzp\") on node \"crc\" DevicePath \"\"" Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.422213 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:26 crc kubenswrapper[4967]: I0120 09:05:26.427659 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8nzf"] Jan 20 09:05:27 crc kubenswrapper[4967]: I0120 09:05:27.701930 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" path="/var/lib/kubelet/pods/3dce87bd-d333-4e97-b31c-c003889108a6/volumes" Jan 20 09:05:27 crc kubenswrapper[4967]: I0120 09:05:27.978251 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-74dcbc6877-8v4vv" Jan 20 09:05:47 crc kubenswrapper[4967]: I0120 09:05:47.474112 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9c99dbd64-q2zrc" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394026 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-96vmf"] Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394475 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394492 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394513 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394525 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394536 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394545 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394556 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394564 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394578 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394585 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394596 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394623 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394637 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394643 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394662 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394668 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="extract-content" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.394677 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394683 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="extract-utilities" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394810 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dce87bd-d333-4e97-b31c-c003889108a6" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394822 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f706bf15-8d26-40a5-8d70-e4a49399009f" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.394836 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="156364ad-181d-441d-9754-104d3a686a3a" containerName="registry-server" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.397224 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.398881 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4"] Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.399980 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.401253 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.401479 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.401834 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.402198 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-t94rh" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.411211 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4"] Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424630 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrxnm\" (UniqueName: \"kubernetes.io/projected/87e441f9-d588-4f98-ad15-d632b239aef5-kube-api-access-wrxnm\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-conf\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics-certs\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424812 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-sockets\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424840 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb28t\" (UniqueName: \"kubernetes.io/projected/0deabec7-6e63-4152-9107-1045d11c8e6c-kube-api-access-nb28t\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424872 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-startup\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.424900 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.425018 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-reloader\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.425088 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/87e441f9-d588-4f98-ad15-d632b239aef5-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.481728 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.481790 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.481836 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.482492 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.482555 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e" gracePeriod=600 Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.514214 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hqtb9"] Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.515113 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.517968 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.518560 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.519251 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.519306 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wfzw6" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526070 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrxnm\" (UniqueName: \"kubernetes.io/projected/87e441f9-d588-4f98-ad15-d632b239aef5-kube-api-access-wrxnm\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526126 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-conf\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526169 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics-certs\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526187 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-sockets\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526206 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb28t\" (UniqueName: \"kubernetes.io/projected/0deabec7-6e63-4152-9107-1045d11c8e6c-kube-api-access-nb28t\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526223 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-startup\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526243 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526294 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-reloader\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.526315 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/87e441f9-d588-4f98-ad15-d632b239aef5-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.527300 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-conf\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.527324 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-sockets\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.527664 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.527921 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0deabec7-6e63-4152-9107-1045d11c8e6c-reloader\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.528463 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0deabec7-6e63-4152-9107-1045d11c8e6c-frr-startup\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.535386 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/87e441f9-d588-4f98-ad15-d632b239aef5-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.547164 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0deabec7-6e63-4152-9107-1045d11c8e6c-metrics-certs\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.574313 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb28t\" (UniqueName: \"kubernetes.io/projected/0deabec7-6e63-4152-9107-1045d11c8e6c-kube-api-access-nb28t\") pod \"frr-k8s-96vmf\" (UID: \"0deabec7-6e63-4152-9107-1045d11c8e6c\") " pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.598673 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrxnm\" (UniqueName: \"kubernetes.io/projected/87e441f9-d588-4f98-ad15-d632b239aef5-kube-api-access-wrxnm\") pod \"frr-k8s-webhook-server-7df86c4f6c-q72g4\" (UID: \"87e441f9-d588-4f98-ad15-d632b239aef5\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.616605 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-gzh9x"] Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.617524 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.622932 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.635969 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdgk\" (UniqueName: \"kubernetes.io/projected/0ac95690-ca20-424c-af78-4134a67a01a2-kube-api-access-bqdgk\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.636031 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.636089 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0ac95690-ca20-424c-af78-4134a67a01a2-metallb-excludel2\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.636120 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.644131 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-gzh9x"] Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.719981 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.726468 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737435 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0ac95690-ca20-424c-af78-4134a67a01a2-metallb-excludel2\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737477 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-cert\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737503 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-metrics-certs\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737525 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737644 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdgk\" (UniqueName: \"kubernetes.io/projected/0ac95690-ca20-424c-af78-4134a67a01a2-kube-api-access-bqdgk\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737743 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.737782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p85ft\" (UniqueName: \"kubernetes.io/projected/180c30fb-af36-47ef-87dc-8d635f6a9e4d-kube-api-access-p85ft\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.738126 4967 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.738162 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0ac95690-ca20-424c-af78-4134a67a01a2-metallb-excludel2\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.738276 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist podName:0ac95690-ca20-424c-af78-4134a67a01a2 nodeName:}" failed. No retries permitted until 2026-01-20 09:05:49.238257901 +0000 UTC m=+983.823558108 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist") pod "speaker-hqtb9" (UID: "0ac95690-ca20-424c-af78-4134a67a01a2") : secret "metallb-memberlist" not found Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.738337 4967 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 20 09:05:48 crc kubenswrapper[4967]: E0120 09:05:48.738369 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs podName:0ac95690-ca20-424c-af78-4134a67a01a2 nodeName:}" failed. No retries permitted until 2026-01-20 09:05:49.238359144 +0000 UTC m=+983.823659461 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs") pod "speaker-hqtb9" (UID: "0ac95690-ca20-424c-af78-4134a67a01a2") : secret "speaker-certs-secret" not found Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.754496 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdgk\" (UniqueName: \"kubernetes.io/projected/0ac95690-ca20-424c-af78-4134a67a01a2-kube-api-access-bqdgk\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.838510 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-metrics-certs\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.838980 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p85ft\" (UniqueName: \"kubernetes.io/projected/180c30fb-af36-47ef-87dc-8d635f6a9e4d-kube-api-access-p85ft\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.839042 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-cert\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.845405 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.846842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-metrics-certs\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.857841 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p85ft\" (UniqueName: \"kubernetes.io/projected/180c30fb-af36-47ef-87dc-8d635f6a9e4d-kube-api-access-p85ft\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.858435 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/180c30fb-af36-47ef-87dc-8d635f6a9e4d-cert\") pod \"controller-6968d8fdc4-gzh9x\" (UID: \"180c30fb-af36-47ef-87dc-8d635f6a9e4d\") " pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.951443 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4"] Jan 20 09:05:48 crc kubenswrapper[4967]: W0120 09:05:48.959875 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e441f9_d588_4f98_ad15_d632b239aef5.slice/crio-058437d2bef185d16fb91399b69083611f18dac8f9cf9420d25b726668a6cedf WatchSource:0}: Error finding container 058437d2bef185d16fb91399b69083611f18dac8f9cf9420d25b726668a6cedf: Status 404 returned error can't find the container with id 058437d2bef185d16fb91399b69083611f18dac8f9cf9420d25b726668a6cedf Jan 20 09:05:48 crc kubenswrapper[4967]: I0120 09:05:48.970095 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.174869 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-gzh9x"] Jan 20 09:05:49 crc kubenswrapper[4967]: W0120 09:05:49.181204 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod180c30fb_af36_47ef_87dc_8d635f6a9e4d.slice/crio-11c319938b88a651c0171d4aab5ff759e5a5eb574d5200aea658f99437c4f043 WatchSource:0}: Error finding container 11c319938b88a651c0171d4aab5ff759e5a5eb574d5200aea658f99437c4f043: Status 404 returned error can't find the container with id 11c319938b88a651c0171d4aab5ff759e5a5eb574d5200aea658f99437c4f043 Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.238739 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-gzh9x" event={"ID":"180c30fb-af36-47ef-87dc-8d635f6a9e4d","Type":"ContainerStarted","Data":"11c319938b88a651c0171d4aab5ff759e5a5eb574d5200aea658f99437c4f043"} Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.239922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" event={"ID":"87e441f9-d588-4f98-ad15-d632b239aef5","Type":"ContainerStarted","Data":"058437d2bef185d16fb91399b69083611f18dac8f9cf9420d25b726668a6cedf"} Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.246300 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.246379 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:49 crc kubenswrapper[4967]: E0120 09:05:49.246511 4967 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 20 09:05:49 crc kubenswrapper[4967]: E0120 09:05:49.246561 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist podName:0ac95690-ca20-424c-af78-4134a67a01a2 nodeName:}" failed. No retries permitted until 2026-01-20 09:05:50.246547652 +0000 UTC m=+984.831847859 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist") pod "speaker-hqtb9" (UID: "0ac95690-ca20-424c-af78-4134a67a01a2") : secret "metallb-memberlist" not found Jan 20 09:05:49 crc kubenswrapper[4967]: I0120 09:05:49.251405 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-metrics-certs\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.246282 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-gzh9x" event={"ID":"180c30fb-af36-47ef-87dc-8d635f6a9e4d","Type":"ContainerStarted","Data":"d5c47bf24cfad96c7880c64e839f8994b33b34e9b72c2caac449b0b56b32afac"} Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.248870 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e" exitCode=0 Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.248895 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e"} Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.248917 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3"} Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.248934 4967 scope.go:117] "RemoveContainer" containerID="2c68cd2d1a402036ef0fbcdb74f0a76f305342220faa8a69c0db8a89a4c8f840" Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.250095 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"4feba0e6346e91b27b2389fe33022d7b18362ad7f121f0743e556c8a2cb493fd"} Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.258539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.265071 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0ac95690-ca20-424c-af78-4134a67a01a2-memberlist\") pod \"speaker-hqtb9\" (UID: \"0ac95690-ca20-424c-af78-4134a67a01a2\") " pod="metallb-system/speaker-hqtb9" Jan 20 09:05:50 crc kubenswrapper[4967]: I0120 09:05:50.333017 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hqtb9" Jan 20 09:05:50 crc kubenswrapper[4967]: W0120 09:05:50.357897 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ac95690_ca20_424c_af78_4134a67a01a2.slice/crio-7e8b704be3d1ec2857382c2f3a5cc6ed926a3b5dcc3fde7c7793f36241866240 WatchSource:0}: Error finding container 7e8b704be3d1ec2857382c2f3a5cc6ed926a3b5dcc3fde7c7793f36241866240: Status 404 returned error can't find the container with id 7e8b704be3d1ec2857382c2f3a5cc6ed926a3b5dcc3fde7c7793f36241866240 Jan 20 09:05:51 crc kubenswrapper[4967]: I0120 09:05:51.255190 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqtb9" event={"ID":"0ac95690-ca20-424c-af78-4134a67a01a2","Type":"ContainerStarted","Data":"6646626290b299a0d8a635ea7e0be385c75c3636aa9bc77c40c004d7a939da1c"} Jan 20 09:05:51 crc kubenswrapper[4967]: I0120 09:05:51.255503 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqtb9" event={"ID":"0ac95690-ca20-424c-af78-4134a67a01a2","Type":"ContainerStarted","Data":"7e8b704be3d1ec2857382c2f3a5cc6ed926a3b5dcc3fde7c7793f36241866240"} Jan 20 09:06:04 crc kubenswrapper[4967]: I0120 09:06:04.374102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-gzh9x" event={"ID":"180c30fb-af36-47ef-87dc-8d635f6a9e4d","Type":"ContainerStarted","Data":"1edd48987eab5bbcfdde1852e6937fb1262ad20455007f83b8f8924928de452a"} Jan 20 09:06:04 crc kubenswrapper[4967]: I0120 09:06:04.374890 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:06:04 crc kubenswrapper[4967]: I0120 09:06:04.380789 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-gzh9x" Jan 20 09:06:04 crc kubenswrapper[4967]: I0120 09:06:04.390010 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-gzh9x" podStartSLOduration=7.117768107 podStartE2EDuration="16.389991223s" podCreationTimestamp="2026-01-20 09:05:48 +0000 UTC" firstStartedPulling="2026-01-20 09:05:49.848128693 +0000 UTC m=+984.433428900" lastFinishedPulling="2026-01-20 09:05:59.120351799 +0000 UTC m=+993.705652016" observedRunningTime="2026-01-20 09:06:04.389123119 +0000 UTC m=+998.974423326" watchObservedRunningTime="2026-01-20 09:06:04.389991223 +0000 UTC m=+998.975291430" Jan 20 09:06:05 crc kubenswrapper[4967]: I0120 09:06:05.384362 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqtb9" event={"ID":"0ac95690-ca20-424c-af78-4134a67a01a2","Type":"ContainerStarted","Data":"791119da83c3b15ee21ef820b9308bda214960761829d6eff4f1126ae367c5ca"} Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.408517 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" event={"ID":"87e441f9-d588-4f98-ad15-d632b239aef5","Type":"ContainerStarted","Data":"94fc7b7dceb5bba377e59faf462deb1b7a225f05fe1ebc9356fc8ab5ccd7ade5"} Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.409871 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.411434 4967 generic.go:334] "Generic (PLEG): container finished" podID="0deabec7-6e63-4152-9107-1045d11c8e6c" containerID="858d86d91e213ca3eea3deee01554289096ba527a245e66efc1aadf170073395" exitCode=0 Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.413278 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerDied","Data":"858d86d91e213ca3eea3deee01554289096ba527a245e66efc1aadf170073395"} Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.413311 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hqtb9" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.435824 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hqtb9" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.470964 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" podStartSLOduration=6.327033124 podStartE2EDuration="21.470947879s" podCreationTimestamp="2026-01-20 09:05:48 +0000 UTC" firstStartedPulling="2026-01-20 09:05:48.966318381 +0000 UTC m=+983.551618588" lastFinishedPulling="2026-01-20 09:06:04.110233136 +0000 UTC m=+998.695533343" observedRunningTime="2026-01-20 09:06:09.435406706 +0000 UTC m=+1004.020706913" watchObservedRunningTime="2026-01-20 09:06:09.470947879 +0000 UTC m=+1004.056248086" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.473730 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hqtb9" podStartSLOduration=13.539747173 podStartE2EDuration="21.473720497s" podCreationTimestamp="2026-01-20 09:05:48 +0000 UTC" firstStartedPulling="2026-01-20 09:05:51.193973689 +0000 UTC m=+985.779273896" lastFinishedPulling="2026-01-20 09:05:59.127947013 +0000 UTC m=+993.713247220" observedRunningTime="2026-01-20 09:06:09.4695974 +0000 UTC m=+1004.054897607" watchObservedRunningTime="2026-01-20 09:06:09.473720497 +0000 UTC m=+1004.059020704" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.958951 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.959637 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:09 crc kubenswrapper[4967]: I0120 09:06:09.965418 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-hxc4c" Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.033258 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.129699 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65wwk\" (UniqueName: \"kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk\") pod \"infra-operator-index-xbtc6\" (UID: \"2b694e6c-00e5-429f-8bd9-8c3e7e035913\") " pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.231291 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65wwk\" (UniqueName: \"kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk\") pod \"infra-operator-index-xbtc6\" (UID: \"2b694e6c-00e5-429f-8bd9-8c3e7e035913\") " pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.249295 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65wwk\" (UniqueName: \"kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk\") pod \"infra-operator-index-xbtc6\" (UID: \"2b694e6c-00e5-429f-8bd9-8c3e7e035913\") " pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.276146 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.422835 4967 generic.go:334] "Generic (PLEG): container finished" podID="0deabec7-6e63-4152-9107-1045d11c8e6c" containerID="cd9b4d9b6bb409e7f27163d4baa8b29d7f608c52ea7548fdfdb40715cdb4fc08" exitCode=0 Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.422901 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerDied","Data":"cd9b4d9b6bb409e7f27163d4baa8b29d7f608c52ea7548fdfdb40715cdb4fc08"} Jan 20 09:06:10 crc kubenswrapper[4967]: I0120 09:06:10.480657 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:10 crc kubenswrapper[4967]: W0120 09:06:10.488829 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b694e6c_00e5_429f_8bd9_8c3e7e035913.slice/crio-cebe23604e733f5af3c775ed9c425c59c645674a9cb4708a72d01cc22fba164f WatchSource:0}: Error finding container cebe23604e733f5af3c775ed9c425c59c645674a9cb4708a72d01cc22fba164f: Status 404 returned error can't find the container with id cebe23604e733f5af3c775ed9c425c59c645674a9cb4708a72d01cc22fba164f Jan 20 09:06:11 crc kubenswrapper[4967]: I0120 09:06:11.436738 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"2afa7a7245110380d538634fde19f46d613aef130a71c935c5e7ca64c57b4427"} Jan 20 09:06:11 crc kubenswrapper[4967]: I0120 09:06:11.439035 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xbtc6" event={"ID":"2b694e6c-00e5-429f-8bd9-8c3e7e035913","Type":"ContainerStarted","Data":"cebe23604e733f5af3c775ed9c425c59c645674a9cb4708a72d01cc22fba164f"} Jan 20 09:06:12 crc kubenswrapper[4967]: I0120 09:06:12.446672 4967 generic.go:334] "Generic (PLEG): container finished" podID="0deabec7-6e63-4152-9107-1045d11c8e6c" containerID="2afa7a7245110380d538634fde19f46d613aef130a71c935c5e7ca64c57b4427" exitCode=0 Jan 20 09:06:12 crc kubenswrapper[4967]: I0120 09:06:12.446745 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerDied","Data":"2afa7a7245110380d538634fde19f46d613aef130a71c935c5e7ca64c57b4427"} Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.137261 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.457152 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xbtc6" event={"ID":"2b694e6c-00e5-429f-8bd9-8c3e7e035913","Type":"ContainerStarted","Data":"7399aa25fc4f7c7216bd0ee25700429db5db1ff96b9fd86060127641a6675f16"} Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.457223 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-xbtc6" podUID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" containerName="registry-server" containerID="cri-o://7399aa25fc4f7c7216bd0ee25700429db5db1ff96b9fd86060127641a6675f16" gracePeriod=2 Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.461096 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"e9ee2653f2402567f9746b873c7cb3ee85503155ca1f0555c7409ebdbc1c7d05"} Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.461385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"a6aa3eab14f8f404e0fded044c1ff2b6d4e3f8b80df24c6e8f294984ba2cb41e"} Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.472829 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-xbtc6" podStartSLOduration=1.721158798 podStartE2EDuration="4.472791618s" podCreationTimestamp="2026-01-20 09:06:09 +0000 UTC" firstStartedPulling="2026-01-20 09:06:10.490918196 +0000 UTC m=+1005.076218403" lastFinishedPulling="2026-01-20 09:06:13.242551016 +0000 UTC m=+1007.827851223" observedRunningTime="2026-01-20 09:06:13.469410753 +0000 UTC m=+1008.054710960" watchObservedRunningTime="2026-01-20 09:06:13.472791618 +0000 UTC m=+1008.058091825" Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.744130 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.745056 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.752648 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.813943 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc9nr\" (UniqueName: \"kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr\") pod \"infra-operator-index-4b7zw\" (UID: \"a1eac7ed-6381-4765-b9a6-1b680641c4c0\") " pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.915507 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc9nr\" (UniqueName: \"kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr\") pod \"infra-operator-index-4b7zw\" (UID: \"a1eac7ed-6381-4765-b9a6-1b680641c4c0\") " pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:13 crc kubenswrapper[4967]: I0120 09:06:13.936435 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc9nr\" (UniqueName: \"kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr\") pod \"infra-operator-index-4b7zw\" (UID: \"a1eac7ed-6381-4765-b9a6-1b680641c4c0\") " pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.126492 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.473102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"516b496116db2914b52e71e9b948954937a3a76ddfe90de0954d8d6441d4ec74"} Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.473462 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"2f2de3fbc174b2e2dfee9cb601ebd09fa1f37cbf69084a60604e2f978fbc0139"} Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.475507 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-xbtc6_2b694e6c-00e5-429f-8bd9-8c3e7e035913/registry-server/0.log" Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.475558 4967 generic.go:334] "Generic (PLEG): container finished" podID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" containerID="7399aa25fc4f7c7216bd0ee25700429db5db1ff96b9fd86060127641a6675f16" exitCode=1 Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.475584 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xbtc6" event={"ID":"2b694e6c-00e5-429f-8bd9-8c3e7e035913","Type":"ContainerDied","Data":"7399aa25fc4f7c7216bd0ee25700429db5db1ff96b9fd86060127641a6675f16"} Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.538360 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:06:14 crc kubenswrapper[4967]: W0120 09:06:14.555461 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1eac7ed_6381_4765_b9a6_1b680641c4c0.slice/crio-01fcbed6d7ae8076502bcfb6c4ee28c040467185f0010b43a3b46caebfc06fc3 WatchSource:0}: Error finding container 01fcbed6d7ae8076502bcfb6c4ee28c040467185f0010b43a3b46caebfc06fc3: Status 404 returned error can't find the container with id 01fcbed6d7ae8076502bcfb6c4ee28c040467185f0010b43a3b46caebfc06fc3 Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.910323 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-xbtc6_2b694e6c-00e5-429f-8bd9-8c3e7e035913/registry-server/0.log" Jan 20 09:06:14 crc kubenswrapper[4967]: I0120 09:06:14.910398 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.036160 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65wwk\" (UniqueName: \"kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk\") pod \"2b694e6c-00e5-429f-8bd9-8c3e7e035913\" (UID: \"2b694e6c-00e5-429f-8bd9-8c3e7e035913\") " Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.043900 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk" (OuterVolumeSpecName: "kube-api-access-65wwk") pod "2b694e6c-00e5-429f-8bd9-8c3e7e035913" (UID: "2b694e6c-00e5-429f-8bd9-8c3e7e035913"). InnerVolumeSpecName "kube-api-access-65wwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.138503 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65wwk\" (UniqueName: \"kubernetes.io/projected/2b694e6c-00e5-429f-8bd9-8c3e7e035913-kube-api-access-65wwk\") on node \"crc\" DevicePath \"\"" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.483290 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4b7zw" event={"ID":"a1eac7ed-6381-4765-b9a6-1b680641c4c0","Type":"ContainerStarted","Data":"01fcbed6d7ae8076502bcfb6c4ee28c040467185f0010b43a3b46caebfc06fc3"} Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.487710 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"4e8e97269558a094abb8281e055f7daae45c22c6ab5ceabb74eec9ee20be64e8"} Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.487758 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96vmf" event={"ID":"0deabec7-6e63-4152-9107-1045d11c8e6c","Type":"ContainerStarted","Data":"2de56df98a87cda6e4ced82f7c47d014f5b38009c73f14b8fda8a4823cc15f35"} Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.487940 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.488824 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-xbtc6_2b694e6c-00e5-429f-8bd9-8c3e7e035913/registry-server/0.log" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.488868 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xbtc6" event={"ID":"2b694e6c-00e5-429f-8bd9-8c3e7e035913","Type":"ContainerDied","Data":"cebe23604e733f5af3c775ed9c425c59c645674a9cb4708a72d01cc22fba164f"} Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.488897 4967 scope.go:117] "RemoveContainer" containerID="7399aa25fc4f7c7216bd0ee25700429db5db1ff96b9fd86060127641a6675f16" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.489006 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xbtc6" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.519839 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-96vmf" podStartSLOduration=12.977977073 podStartE2EDuration="27.519820123s" podCreationTimestamp="2026-01-20 09:05:48 +0000 UTC" firstStartedPulling="2026-01-20 09:05:49.523823739 +0000 UTC m=+984.109123946" lastFinishedPulling="2026-01-20 09:06:04.065666789 +0000 UTC m=+998.650966996" observedRunningTime="2026-01-20 09:06:15.516043827 +0000 UTC m=+1010.101344074" watchObservedRunningTime="2026-01-20 09:06:15.519820123 +0000 UTC m=+1010.105120330" Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.531283 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.536332 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-xbtc6"] Jan 20 09:06:15 crc kubenswrapper[4967]: I0120 09:06:15.703861 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" path="/var/lib/kubelet/pods/2b694e6c-00e5-429f-8bd9-8c3e7e035913/volumes" Jan 20 09:06:18 crc kubenswrapper[4967]: I0120 09:06:18.721041 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:06:18 crc kubenswrapper[4967]: I0120 09:06:18.763208 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:06:21 crc kubenswrapper[4967]: I0120 09:06:21.526889 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4b7zw" event={"ID":"a1eac7ed-6381-4765-b9a6-1b680641c4c0","Type":"ContainerStarted","Data":"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f"} Jan 20 09:06:21 crc kubenswrapper[4967]: I0120 09:06:21.549673 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-4b7zw" podStartSLOduration=2.3867649220000002 podStartE2EDuration="8.5496073s" podCreationTimestamp="2026-01-20 09:06:13 +0000 UTC" firstStartedPulling="2026-01-20 09:06:14.56243925 +0000 UTC m=+1009.147739457" lastFinishedPulling="2026-01-20 09:06:20.725281608 +0000 UTC m=+1015.310581835" observedRunningTime="2026-01-20 09:06:21.538989271 +0000 UTC m=+1016.124289498" watchObservedRunningTime="2026-01-20 09:06:21.5496073 +0000 UTC m=+1016.134907547" Jan 20 09:06:22 crc kubenswrapper[4967]: I0120 09:06:22.533027 4967 generic.go:334] "Generic (PLEG): container finished" podID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerID="8946ebfdbb11e9d1b449ffdce1dcc75cdf64bec416e1d16252aaf05ccec4987f" exitCode=1 Jan 20 09:06:22 crc kubenswrapper[4967]: I0120 09:06:22.533147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerDied","Data":"8946ebfdbb11e9d1b449ffdce1dcc75cdf64bec416e1d16252aaf05ccec4987f"} Jan 20 09:06:22 crc kubenswrapper[4967]: I0120 09:06:22.533725 4967 scope.go:117] "RemoveContainer" containerID="8946ebfdbb11e9d1b449ffdce1dcc75cdf64bec416e1d16252aaf05ccec4987f" Jan 20 09:06:23 crc kubenswrapper[4967]: I0120 09:06:23.540699 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerStarted","Data":"706826535e4a9cc773ec188a6496d0f3c6ced4eca112d3293f987ea14d53a4f6"} Jan 20 09:06:23 crc kubenswrapper[4967]: I0120 09:06:23.541210 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:06:24 crc kubenswrapper[4967]: I0120 09:06:24.127456 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:24 crc kubenswrapper[4967]: I0120 09:06:24.127518 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:24 crc kubenswrapper[4967]: I0120 09:06:24.153407 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:28 crc kubenswrapper[4967]: I0120 09:06:28.728705 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-96vmf" Jan 20 09:06:28 crc kubenswrapper[4967]: I0120 09:06:28.734991 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-q72g4" Jan 20 09:06:34 crc kubenswrapper[4967]: I0120 09:06:34.157439 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:06:34 crc kubenswrapper[4967]: I0120 09:06:34.674849 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.986913 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9"] Jan 20 09:06:44 crc kubenswrapper[4967]: E0120 09:06:44.987567 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" containerName="registry-server" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.987584 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" containerName="registry-server" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.987755 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b694e6c-00e5-429f-8bd9-8c3e7e035913" containerName="registry-server" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.988558 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.990269 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lqxxp" Jan 20 09:06:44 crc kubenswrapper[4967]: I0120 09:06:44.995469 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9"] Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.108391 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.108443 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfwgc\" (UniqueName: \"kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.108532 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.209350 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.210075 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfwgc\" (UniqueName: \"kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.210175 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.210299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.210591 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.235031 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfwgc\" (UniqueName: \"kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc\") pod \"ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.352363 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.563897 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9"] Jan 20 09:06:45 crc kubenswrapper[4967]: W0120 09:06:45.566485 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd010afb9_5722_465f_a06d_bd3edaacc210.slice/crio-3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4 WatchSource:0}: Error finding container 3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4: Status 404 returned error can't find the container with id 3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4 Jan 20 09:06:45 crc kubenswrapper[4967]: I0120 09:06:45.669437 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" event={"ID":"d010afb9-5722-465f-a06d-bd3edaacc210","Type":"ContainerStarted","Data":"3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4"} Jan 20 09:06:46 crc kubenswrapper[4967]: I0120 09:06:46.677800 4967 generic.go:334] "Generic (PLEG): container finished" podID="d010afb9-5722-465f-a06d-bd3edaacc210" containerID="99254b5141c34acb9f76f0723e298d765c9033b7f9ba283efa5b225b46291a7f" exitCode=0 Jan 20 09:06:46 crc kubenswrapper[4967]: I0120 09:06:46.677857 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" event={"ID":"d010afb9-5722-465f-a06d-bd3edaacc210","Type":"ContainerDied","Data":"99254b5141c34acb9f76f0723e298d765c9033b7f9ba283efa5b225b46291a7f"} Jan 20 09:06:51 crc kubenswrapper[4967]: I0120 09:06:51.708262 4967 generic.go:334] "Generic (PLEG): container finished" podID="d010afb9-5722-465f-a06d-bd3edaacc210" containerID="839283610dc65f6081a084c2ecf836fc16beb1b7f9a03d6eeb8a0875e5196b8d" exitCode=0 Jan 20 09:06:51 crc kubenswrapper[4967]: I0120 09:06:51.708315 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" event={"ID":"d010afb9-5722-465f-a06d-bd3edaacc210","Type":"ContainerDied","Data":"839283610dc65f6081a084c2ecf836fc16beb1b7f9a03d6eeb8a0875e5196b8d"} Jan 20 09:06:52 crc kubenswrapper[4967]: I0120 09:06:52.717127 4967 generic.go:334] "Generic (PLEG): container finished" podID="d010afb9-5722-465f-a06d-bd3edaacc210" containerID="e47033e9c8290a2c0e606c58e8a59d063f0252cad126658b578380d065358ce4" exitCode=0 Jan 20 09:06:52 crc kubenswrapper[4967]: I0120 09:06:52.717195 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" event={"ID":"d010afb9-5722-465f-a06d-bd3edaacc210","Type":"ContainerDied","Data":"e47033e9c8290a2c0e606c58e8a59d063f0252cad126658b578380d065358ce4"} Jan 20 09:06:53 crc kubenswrapper[4967]: I0120 09:06:53.981771 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.036144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfwgc\" (UniqueName: \"kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc\") pod \"d010afb9-5722-465f-a06d-bd3edaacc210\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.036280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle\") pod \"d010afb9-5722-465f-a06d-bd3edaacc210\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.036321 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util\") pod \"d010afb9-5722-465f-a06d-bd3edaacc210\" (UID: \"d010afb9-5722-465f-a06d-bd3edaacc210\") " Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.038690 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle" (OuterVolumeSpecName: "bundle") pod "d010afb9-5722-465f-a06d-bd3edaacc210" (UID: "d010afb9-5722-465f-a06d-bd3edaacc210"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.043170 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc" (OuterVolumeSpecName: "kube-api-access-zfwgc") pod "d010afb9-5722-465f-a06d-bd3edaacc210" (UID: "d010afb9-5722-465f-a06d-bd3edaacc210"). InnerVolumeSpecName "kube-api-access-zfwgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.048651 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util" (OuterVolumeSpecName: "util") pod "d010afb9-5722-465f-a06d-bd3edaacc210" (UID: "d010afb9-5722-465f-a06d-bd3edaacc210"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.137711 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.137744 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d010afb9-5722-465f-a06d-bd3edaacc210-util\") on node \"crc\" DevicePath \"\"" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.137754 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfwgc\" (UniqueName: \"kubernetes.io/projected/d010afb9-5722-465f-a06d-bd3edaacc210-kube-api-access-zfwgc\") on node \"crc\" DevicePath \"\"" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.731216 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" event={"ID":"d010afb9-5722-465f-a06d-bd3edaacc210","Type":"ContainerDied","Data":"3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4"} Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.731535 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d7981145bbf928af9c74d0db44c9454240f494ef8d9d1a9515dc756836d7dc4" Jan 20 09:06:54 crc kubenswrapper[4967]: I0120 09:06:54.731264 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.563687 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:07:03 crc kubenswrapper[4967]: E0120 09:07:03.564376 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="util" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.564387 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="util" Jan 20 09:07:03 crc kubenswrapper[4967]: E0120 09:07:03.564401 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="extract" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.564407 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="extract" Jan 20 09:07:03 crc kubenswrapper[4967]: E0120 09:07:03.564421 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="pull" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.564427 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="pull" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.564523 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" containerName="extract" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.564956 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.566721 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-h6p79" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.567328 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.582557 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.765594 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.765797 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.765896 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp8tc\" (UniqueName: \"kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.866915 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp8tc\" (UniqueName: \"kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.867014 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.867066 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.873668 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.875724 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:03 crc kubenswrapper[4967]: I0120 09:07:03.900503 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp8tc\" (UniqueName: \"kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc\") pod \"infra-operator-controller-manager-6f9888886b-cbv7g\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:04 crc kubenswrapper[4967]: I0120 09:07:04.180770 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:04 crc kubenswrapper[4967]: I0120 09:07:04.363751 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:07:04 crc kubenswrapper[4967]: I0120 09:07:04.789401 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" event={"ID":"a4fc08d9-6214-48db-9f0c-38cda7310168","Type":"ContainerStarted","Data":"45f10f613d721e069edb97db9c86894cee5c69028143fa77ca0d7c3720f70296"} Jan 20 09:07:07 crc kubenswrapper[4967]: I0120 09:07:07.815323 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" event={"ID":"a4fc08d9-6214-48db-9f0c-38cda7310168","Type":"ContainerStarted","Data":"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c"} Jan 20 09:07:07 crc kubenswrapper[4967]: I0120 09:07:07.815867 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:07 crc kubenswrapper[4967]: I0120 09:07:07.840407 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" podStartSLOduration=2.037430483 podStartE2EDuration="4.84039067s" podCreationTimestamp="2026-01-20 09:07:03 +0000 UTC" firstStartedPulling="2026-01-20 09:07:04.377532312 +0000 UTC m=+1058.962832519" lastFinishedPulling="2026-01-20 09:07:07.180492499 +0000 UTC m=+1061.765792706" observedRunningTime="2026-01-20 09:07:07.838188759 +0000 UTC m=+1062.423488966" watchObservedRunningTime="2026-01-20 09:07:07.84039067 +0000 UTC m=+1062.425690877" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.050184 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.051399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.057984 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"kube-root-ca.crt" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.057992 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-config-data" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.058238 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openshift-service-ca.crt" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.058304 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-scripts" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.061357 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"galera-openstack-dockercfg-78bmf" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.065389 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.066777 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.088428 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.089730 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.093338 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.107202 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.149772 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163423 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163469 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163492 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163514 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163657 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xshsd\" (UniqueName: \"kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.163726 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264595 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2895c\" (UniqueName: \"kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264730 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264762 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264784 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264811 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264835 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264864 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264883 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xshsd\" (UniqueName: \"kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264929 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264954 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.264980 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265070 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265136 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65kbt\" (UniqueName: \"kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265189 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265260 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265373 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") device mount path \"/mnt/openstack/pv05\"" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265424 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265666 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265667 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.265840 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.267094 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.292034 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.296670 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xshsd\" (UniqueName: \"kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd\") pod \"openstack-galera-0\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367230 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65kbt\" (UniqueName: \"kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367302 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367337 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367407 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367447 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2895c\" (UniqueName: \"kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367484 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367517 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367575 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367600 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367672 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367817 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") device mount path \"/mnt/openstack/pv03\"" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367837 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") device mount path \"/mnt/openstack/pv07\"" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.367899 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.368087 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.368401 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.368428 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.368681 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.368820 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.369086 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.370596 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.373864 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.383047 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.384270 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65kbt\" (UniqueName: \"kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.389328 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2895c\" (UniqueName: \"kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c\") pod \"openstack-galera-1\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.390345 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-2\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.404961 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.659543 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:07:10 crc kubenswrapper[4967]: W0120 09:07:10.665931 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod689be023_b3b2_4b2c_b16e_5a24ce36edfc.slice/crio-375d2ac1b5cd2c9bd022f3db354145e25ce423f089e43fa09e9cd0ca66e560c6 WatchSource:0}: Error finding container 375d2ac1b5cd2c9bd022f3db354145e25ce423f089e43fa09e9cd0ca66e560c6: Status 404 returned error can't find the container with id 375d2ac1b5cd2c9bd022f3db354145e25ce423f089e43fa09e9cd0ca66e560c6 Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.685573 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.811931 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:07:10 crc kubenswrapper[4967]: W0120 09:07:10.830002 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdd3d19e_4ca0_4a64_9a2d_a7dd13f8be22.slice/crio-7ea6cf633ff8f5d66a0bb2857d37b1efbb3f36a46e7c2e67570616c2913482ea WatchSource:0}: Error finding container 7ea6cf633ff8f5d66a0bb2857d37b1efbb3f36a46e7c2e67570616c2913482ea: Status 404 returned error can't find the container with id 7ea6cf633ff8f5d66a0bb2857d37b1efbb3f36a46e7c2e67570616c2913482ea Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.831523 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerStarted","Data":"375d2ac1b5cd2c9bd022f3db354145e25ce423f089e43fa09e9cd0ca66e560c6"} Jan 20 09:07:10 crc kubenswrapper[4967]: I0120 09:07:10.903265 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:07:11 crc kubenswrapper[4967]: I0120 09:07:11.847202 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerStarted","Data":"0a9eeec82f86eb69bfd3bdd261a62bc1f82e89e2f24d4118d7f080b8d358f4aa"} Jan 20 09:07:11 crc kubenswrapper[4967]: I0120 09:07:11.848576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerStarted","Data":"7ea6cf633ff8f5d66a0bb2857d37b1efbb3f36a46e7c2e67570616c2913482ea"} Jan 20 09:07:14 crc kubenswrapper[4967]: I0120 09:07:14.205449 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.717264 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.719090 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.730716 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"memcached-config-data" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.730885 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"memcached-memcached-dockercfg-j5rc6" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.745981 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.851326 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.851683 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ppzp\" (UniqueName: \"kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.851798 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.953282 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.953361 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ppzp\" (UniqueName: \"kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.953394 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.954261 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.954963 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:15 crc kubenswrapper[4967]: I0120 09:07:15.977902 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ppzp\" (UniqueName: \"kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp\") pod \"memcached-0\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.038954 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.305650 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.700573 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.701631 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.704797 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-lqnch" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.713932 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.869702 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f8dq\" (UniqueName: \"kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq\") pod \"rabbitmq-cluster-operator-index-cqvc9\" (UID: \"c482d01f-03f6-4401-b740-1f4a804ed359\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.882487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"5030695f-a34f-47ef-acaa-c0dfa3601fd8","Type":"ContainerStarted","Data":"1a1599c9a7c37ad6a43b76420cd8dba7b91204b3fa338df24093b5ddd762aa13"} Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.971081 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f8dq\" (UniqueName: \"kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq\") pod \"rabbitmq-cluster-operator-index-cqvc9\" (UID: \"c482d01f-03f6-4401-b740-1f4a804ed359\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:16 crc kubenswrapper[4967]: I0120 09:07:16.992039 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f8dq\" (UniqueName: \"kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq\") pod \"rabbitmq-cluster-operator-index-cqvc9\" (UID: \"c482d01f-03f6-4401-b740-1f4a804ed359\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:17 crc kubenswrapper[4967]: I0120 09:07:17.023767 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:20 crc kubenswrapper[4967]: I0120 09:07:20.888621 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.494911 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.496270 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.503910 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.635209 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxkpj\" (UniqueName: \"kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj\") pod \"rabbitmq-cluster-operator-index-qsnbm\" (UID: \"82afd249-e563-437f-9011-05be3f663020\") " pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.737345 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxkpj\" (UniqueName: \"kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj\") pod \"rabbitmq-cluster-operator-index-qsnbm\" (UID: \"82afd249-e563-437f-9011-05be3f663020\") " pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.760680 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxkpj\" (UniqueName: \"kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj\") pod \"rabbitmq-cluster-operator-index-qsnbm\" (UID: \"82afd249-e563-437f-9011-05be3f663020\") " pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:21 crc kubenswrapper[4967]: I0120 09:07:21.821369 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:25 crc kubenswrapper[4967]: I0120 09:07:25.104686 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:07:25 crc kubenswrapper[4967]: I0120 09:07:25.220629 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:26 crc kubenswrapper[4967]: I0120 09:07:26.027258 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" event={"ID":"c482d01f-03f6-4401-b740-1f4a804ed359","Type":"ContainerStarted","Data":"43048f08e3fc6f20107430f6e7c1cd9ff8aa78bbdf6305fd1361325f0823d25a"} Jan 20 09:07:26 crc kubenswrapper[4967]: I0120 09:07:26.030430 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" event={"ID":"82afd249-e563-437f-9011-05be3f663020","Type":"ContainerStarted","Data":"fa99a32ca18dbe687ccddac6930948876d537078aac2d861c07b95b832bf11d3"} Jan 20 09:07:27 crc kubenswrapper[4967]: I0120 09:07:27.039780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerStarted","Data":"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9"} Jan 20 09:07:27 crc kubenswrapper[4967]: I0120 09:07:27.042332 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerStarted","Data":"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0"} Jan 20 09:07:27 crc kubenswrapper[4967]: I0120 09:07:27.044810 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerStarted","Data":"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7"} Jan 20 09:07:37 crc kubenswrapper[4967]: I0120 09:07:37.103301 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"5030695f-a34f-47ef-acaa-c0dfa3601fd8","Type":"ContainerStarted","Data":"5a745e943b60dccc0dcba24f62de592b210fe6ab152e694350a3f95b818fccf9"} Jan 20 09:07:39 crc kubenswrapper[4967]: I0120 09:07:39.113362 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:39 crc kubenswrapper[4967]: I0120 09:07:39.132498 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/memcached-0" podStartSLOduration=3.793795449 podStartE2EDuration="24.132476879s" podCreationTimestamp="2026-01-20 09:07:15 +0000 UTC" firstStartedPulling="2026-01-20 09:07:16.316360701 +0000 UTC m=+1070.901660898" lastFinishedPulling="2026-01-20 09:07:36.655042131 +0000 UTC m=+1091.240342328" observedRunningTime="2026-01-20 09:07:39.129465356 +0000 UTC m=+1093.714765583" watchObservedRunningTime="2026-01-20 09:07:39.132476879 +0000 UTC m=+1093.717777086" Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.040882 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.131944 4967 generic.go:334] "Generic (PLEG): container finished" podID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerID="8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0" exitCode=0 Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.132044 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerDied","Data":"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0"} Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.136334 4967 generic.go:334] "Generic (PLEG): container finished" podID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerID="2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7" exitCode=0 Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.136465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerDied","Data":"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7"} Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.139249 4967 generic.go:334] "Generic (PLEG): container finished" podID="7ce51cd1-0945-4880-8571-3779909444d6" containerID="598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9" exitCode=0 Jan 20 09:07:41 crc kubenswrapper[4967]: I0120 09:07:41.139347 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerDied","Data":"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9"} Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.163845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerStarted","Data":"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92"} Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.166991 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerStarted","Data":"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d"} Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.169876 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerStarted","Data":"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0"} Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.196336 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-2" podStartSLOduration=21.100939716 podStartE2EDuration="36.196312952s" podCreationTimestamp="2026-01-20 09:07:09 +0000 UTC" firstStartedPulling="2026-01-20 09:07:10.66954683 +0000 UTC m=+1065.254847037" lastFinishedPulling="2026-01-20 09:07:25.764920066 +0000 UTC m=+1080.350220273" observedRunningTime="2026-01-20 09:07:45.188069535 +0000 UTC m=+1099.773369752" watchObservedRunningTime="2026-01-20 09:07:45.196312952 +0000 UTC m=+1099.781613179" Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.222502 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-0" podStartSLOduration=21.449099099 podStartE2EDuration="36.222482674s" podCreationTimestamp="2026-01-20 09:07:09 +0000 UTC" firstStartedPulling="2026-01-20 09:07:10.831914198 +0000 UTC m=+1065.417214405" lastFinishedPulling="2026-01-20 09:07:25.605297773 +0000 UTC m=+1080.190597980" observedRunningTime="2026-01-20 09:07:45.221861867 +0000 UTC m=+1099.807162084" watchObservedRunningTime="2026-01-20 09:07:45.222482674 +0000 UTC m=+1099.807782891" Jan 20 09:07:45 crc kubenswrapper[4967]: I0120 09:07:45.244766 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstack-galera-1" podStartSLOduration=21.472905495 podStartE2EDuration="36.244747628s" podCreationTimestamp="2026-01-20 09:07:09 +0000 UTC" firstStartedPulling="2026-01-20 09:07:10.914931947 +0000 UTC m=+1065.500232154" lastFinishedPulling="2026-01-20 09:07:25.68677408 +0000 UTC m=+1080.272074287" observedRunningTime="2026-01-20 09:07:45.241641112 +0000 UTC m=+1099.826941389" watchObservedRunningTime="2026-01-20 09:07:45.244747628 +0000 UTC m=+1099.830047845" Jan 20 09:07:47 crc kubenswrapper[4967]: I0120 09:07:47.182180 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" event={"ID":"82afd249-e563-437f-9011-05be3f663020","Type":"ContainerStarted","Data":"2eec8faabd74788596b928356698a73692a061101b4c0e23b10f2ad38c549cde"} Jan 20 09:07:48 crc kubenswrapper[4967]: I0120 09:07:48.189033 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" event={"ID":"c482d01f-03f6-4401-b740-1f4a804ed359","Type":"ContainerStarted","Data":"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff"} Jan 20 09:07:48 crc kubenswrapper[4967]: I0120 09:07:48.189260 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" podUID="c482d01f-03f6-4401-b740-1f4a804ed359" containerName="registry-server" containerID="cri-o://318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff" gracePeriod=2 Jan 20 09:07:48 crc kubenswrapper[4967]: I0120 09:07:48.207765 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" podStartSLOduration=10.828394549 podStartE2EDuration="32.207746489s" podCreationTimestamp="2026-01-20 09:07:16 +0000 UTC" firstStartedPulling="2026-01-20 09:07:25.224967465 +0000 UTC m=+1079.810267672" lastFinishedPulling="2026-01-20 09:07:46.604319405 +0000 UTC m=+1101.189619612" observedRunningTime="2026-01-20 09:07:48.205025154 +0000 UTC m=+1102.790325361" watchObservedRunningTime="2026-01-20 09:07:48.207746489 +0000 UTC m=+1102.793046696" Jan 20 09:07:48 crc kubenswrapper[4967]: I0120 09:07:48.208111 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" podStartSLOduration=6.324773097 podStartE2EDuration="27.208105439s" podCreationTimestamp="2026-01-20 09:07:21 +0000 UTC" firstStartedPulling="2026-01-20 09:07:25.119750532 +0000 UTC m=+1079.705050739" lastFinishedPulling="2026-01-20 09:07:46.003082874 +0000 UTC m=+1100.588383081" observedRunningTime="2026-01-20 09:07:47.249470099 +0000 UTC m=+1101.834770316" watchObservedRunningTime="2026-01-20 09:07:48.208105439 +0000 UTC m=+1102.793405646" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.151272 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.194104 4967 generic.go:334] "Generic (PLEG): container finished" podID="c482d01f-03f6-4401-b740-1f4a804ed359" containerID="318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff" exitCode=0 Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.194140 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" event={"ID":"c482d01f-03f6-4401-b740-1f4a804ed359","Type":"ContainerDied","Data":"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff"} Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.194162 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" event={"ID":"c482d01f-03f6-4401-b740-1f4a804ed359","Type":"ContainerDied","Data":"43048f08e3fc6f20107430f6e7c1cd9ff8aa78bbdf6305fd1361325f0823d25a"} Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.194177 4967 scope.go:117] "RemoveContainer" containerID="318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.194270 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cqvc9" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.211425 4967 scope.go:117] "RemoveContainer" containerID="318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff" Jan 20 09:07:49 crc kubenswrapper[4967]: E0120 09:07:49.211930 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff\": container with ID starting with 318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff not found: ID does not exist" containerID="318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.211961 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff"} err="failed to get container status \"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff\": rpc error: code = NotFound desc = could not find container \"318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff\": container with ID starting with 318b097029841b0d78d2a118b9bc924f5f284ab28a6be0429e2ffbf95781b6ff not found: ID does not exist" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.277863 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f8dq\" (UniqueName: \"kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq\") pod \"c482d01f-03f6-4401-b740-1f4a804ed359\" (UID: \"c482d01f-03f6-4401-b740-1f4a804ed359\") " Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.283851 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq" (OuterVolumeSpecName: "kube-api-access-2f8dq") pod "c482d01f-03f6-4401-b740-1f4a804ed359" (UID: "c482d01f-03f6-4401-b740-1f4a804ed359"). InnerVolumeSpecName "kube-api-access-2f8dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.378938 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f8dq\" (UniqueName: \"kubernetes.io/projected/c482d01f-03f6-4401-b740-1f4a804ed359-kube-api-access-2f8dq\") on node \"crc\" DevicePath \"\"" Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.520560 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.524765 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cqvc9"] Jan 20 09:07:49 crc kubenswrapper[4967]: I0120 09:07:49.714326 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c482d01f-03f6-4401-b740-1f4a804ed359" path="/var/lib/kubelet/pods/c482d01f-03f6-4401-b740-1f4a804ed359/volumes" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.374797 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.374901 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.405255 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.405634 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.686397 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:50 crc kubenswrapper[4967]: I0120 09:07:50.686482 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:07:51 crc kubenswrapper[4967]: I0120 09:07:51.822411 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:51 crc kubenswrapper[4967]: I0120 09:07:51.822904 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:51 crc kubenswrapper[4967]: I0120 09:07:51.853636 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:07:52 crc kubenswrapper[4967]: I0120 09:07:52.243586 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.128276 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8"] Jan 20 09:08:04 crc kubenswrapper[4967]: E0120 09:08:04.129039 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c482d01f-03f6-4401-b740-1f4a804ed359" containerName="registry-server" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.129056 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="c482d01f-03f6-4401-b740-1f4a804ed359" containerName="registry-server" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.129204 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="c482d01f-03f6-4401-b740-1f4a804ed359" containerName="registry-server" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.130293 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.134661 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lqxxp" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.136070 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8"] Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.186819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.187004 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.187035 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hcl8\" (UniqueName: \"kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.288790 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.288837 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hcl8\" (UniqueName: \"kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.288889 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.289355 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.289382 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.311398 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hcl8\" (UniqueName: \"kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.452992 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.543530 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.629233 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:08:04 crc kubenswrapper[4967]: E0120 09:08:04.837445 4967 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.136:39118->38.102.83.136:46175: read tcp 38.102.83.136:39118->38.102.83.136:46175: read: connection reset by peer Jan 20 09:08:04 crc kubenswrapper[4967]: I0120 09:08:04.855559 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8"] Jan 20 09:08:05 crc kubenswrapper[4967]: I0120 09:08:05.288155 4967 generic.go:334] "Generic (PLEG): container finished" podID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerID="2c4816b994da454fa5dabde287637959a4064f487a65bcc70c63892942a32dd6" exitCode=0 Jan 20 09:08:05 crc kubenswrapper[4967]: I0120 09:08:05.288276 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" event={"ID":"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2","Type":"ContainerDied","Data":"2c4816b994da454fa5dabde287637959a4064f487a65bcc70c63892942a32dd6"} Jan 20 09:08:05 crc kubenswrapper[4967]: I0120 09:08:05.288359 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" event={"ID":"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2","Type":"ContainerStarted","Data":"cbb04b9d219df722dba45108b9805e2dcb40bef810781967361b16f49493f2a2"} Jan 20 09:08:08 crc kubenswrapper[4967]: I0120 09:08:08.316465 4967 generic.go:334] "Generic (PLEG): container finished" podID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerID="bf758abfee8d25dd67d8d7924ed8eb925e4aa8e06c57021a1fe24a3aff461a48" exitCode=0 Jan 20 09:08:08 crc kubenswrapper[4967]: I0120 09:08:08.317922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" event={"ID":"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2","Type":"ContainerDied","Data":"bf758abfee8d25dd67d8d7924ed8eb925e4aa8e06c57021a1fe24a3aff461a48"} Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.114838 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/root-account-create-update-jffl5"] Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.116366 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.120074 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"openstack-mariadb-root-db-secret" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.124680 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-jffl5"] Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.173907 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.174156 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd64p\" (UniqueName: \"kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.275516 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd64p\" (UniqueName: \"kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.275572 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.276305 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.303906 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd64p\" (UniqueName: \"kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p\") pod \"root-account-create-update-jffl5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.328850 4967 generic.go:334] "Generic (PLEG): container finished" podID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerID="7bd938a5ab03de0a27d4c58531d093380d90cdfa7b4aed9834e1c842c06b7009" exitCode=0 Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.329157 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" event={"ID":"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2","Type":"ContainerDied","Data":"7bd938a5ab03de0a27d4c58531d093380d90cdfa7b4aed9834e1c842c06b7009"} Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.433349 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:09 crc kubenswrapper[4967]: I0120 09:08:09.855223 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-jffl5"] Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.337573 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/root-account-create-update-jffl5" event={"ID":"135dd5ea-bbbd-4412-a188-8725e5dc8ee5","Type":"ContainerStarted","Data":"5e764437a596ed3a0227ac146257586716ffabeb38e8253b06740f511b6b411a"} Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.337643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/root-account-create-update-jffl5" event={"ID":"135dd5ea-bbbd-4412-a188-8725e5dc8ee5","Type":"ContainerStarted","Data":"e9dfec6e916407234e75d9a40dd04699ca6d66d3a09d28d79145ddbf6067e61d"} Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.356951 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/root-account-create-update-jffl5" podStartSLOduration=1.356930213 podStartE2EDuration="1.356930213s" podCreationTimestamp="2026-01-20 09:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:08:10.353989432 +0000 UTC m=+1124.939289639" watchObservedRunningTime="2026-01-20 09:08:10.356930213 +0000 UTC m=+1124.942230420" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.480724 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="keystone-kuttl-tests/openstack-galera-2" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="galera" probeResult="failure" output=< Jan 20 09:08:10 crc kubenswrapper[4967]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Jan 20 09:08:10 crc kubenswrapper[4967]: > Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.570557 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.693000 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util\") pod \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.693169 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hcl8\" (UniqueName: \"kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8\") pod \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.693207 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle\") pod \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\" (UID: \"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2\") " Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.694142 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle" (OuterVolumeSpecName: "bundle") pod "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" (UID: "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.701933 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8" (OuterVolumeSpecName: "kube-api-access-5hcl8") pod "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" (UID: "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2"). InnerVolumeSpecName "kube-api-access-5hcl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.703586 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util" (OuterVolumeSpecName: "util") pod "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" (UID: "cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.794865 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hcl8\" (UniqueName: \"kubernetes.io/projected/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-kube-api-access-5hcl8\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.794898 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:10 crc kubenswrapper[4967]: I0120 09:08:10.794908 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2-util\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:11 crc kubenswrapper[4967]: I0120 09:08:11.354485 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" event={"ID":"cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2","Type":"ContainerDied","Data":"cbb04b9d219df722dba45108b9805e2dcb40bef810781967361b16f49493f2a2"} Jan 20 09:08:11 crc kubenswrapper[4967]: I0120 09:08:11.354539 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbb04b9d219df722dba45108b9805e2dcb40bef810781967361b16f49493f2a2" Jan 20 09:08:11 crc kubenswrapper[4967]: I0120 09:08:11.354741 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8" Jan 20 09:08:12 crc kubenswrapper[4967]: I0120 09:08:12.361066 4967 generic.go:334] "Generic (PLEG): container finished" podID="135dd5ea-bbbd-4412-a188-8725e5dc8ee5" containerID="5e764437a596ed3a0227ac146257586716ffabeb38e8253b06740f511b6b411a" exitCode=0 Jan 20 09:08:12 crc kubenswrapper[4967]: I0120 09:08:12.361139 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/root-account-create-update-jffl5" event={"ID":"135dd5ea-bbbd-4412-a188-8725e5dc8ee5","Type":"ContainerDied","Data":"5e764437a596ed3a0227ac146257586716ffabeb38e8253b06740f511b6b411a"} Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.643415 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.733809 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd64p\" (UniqueName: \"kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p\") pod \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.733867 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts\") pod \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\" (UID: \"135dd5ea-bbbd-4412-a188-8725e5dc8ee5\") " Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.734688 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "135dd5ea-bbbd-4412-a188-8725e5dc8ee5" (UID: "135dd5ea-bbbd-4412-a188-8725e5dc8ee5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.739355 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p" (OuterVolumeSpecName: "kube-api-access-zd64p") pod "135dd5ea-bbbd-4412-a188-8725e5dc8ee5" (UID: "135dd5ea-bbbd-4412-a188-8725e5dc8ee5"). InnerVolumeSpecName "kube-api-access-zd64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.835141 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd64p\" (UniqueName: \"kubernetes.io/projected/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-kube-api-access-zd64p\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:13 crc kubenswrapper[4967]: I0120 09:08:13.835176 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/135dd5ea-bbbd-4412-a188-8725e5dc8ee5-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:14 crc kubenswrapper[4967]: I0120 09:08:14.016667 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:08:14 crc kubenswrapper[4967]: I0120 09:08:14.122300 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:08:14 crc kubenswrapper[4967]: I0120 09:08:14.375566 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-jffl5" Jan 20 09:08:14 crc kubenswrapper[4967]: I0120 09:08:14.375593 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/root-account-create-update-jffl5" event={"ID":"135dd5ea-bbbd-4412-a188-8725e5dc8ee5","Type":"ContainerDied","Data":"e9dfec6e916407234e75d9a40dd04699ca6d66d3a09d28d79145ddbf6067e61d"} Jan 20 09:08:14 crc kubenswrapper[4967]: I0120 09:08:14.375654 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9dfec6e916407234e75d9a40dd04699ca6d66d3a09d28d79145ddbf6067e61d" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453178 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:08:17 crc kubenswrapper[4967]: E0120 09:08:17.453461 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="pull" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453475 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="pull" Jan 20 09:08:17 crc kubenswrapper[4967]: E0120 09:08:17.453494 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="util" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453502 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="util" Jan 20 09:08:17 crc kubenswrapper[4967]: E0120 09:08:17.453520 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="extract" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453527 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="extract" Jan 20 09:08:17 crc kubenswrapper[4967]: E0120 09:08:17.453538 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135dd5ea-bbbd-4412-a188-8725e5dc8ee5" containerName="mariadb-account-create-update" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453547 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="135dd5ea-bbbd-4412-a188-8725e5dc8ee5" containerName="mariadb-account-create-update" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453699 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" containerName="extract" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.453719 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="135dd5ea-bbbd-4412-a188-8725e5dc8ee5" containerName="mariadb-account-create-update" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.454194 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.458807 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-v2rxr" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.463567 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.479157 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwlr\" (UniqueName: \"kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr\") pod \"rabbitmq-cluster-operator-779fc9694b-nzkln\" (UID: \"40c1982c-ddc0-4043-a956-2a5332612176\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.580313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwlr\" (UniqueName: \"kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr\") pod \"rabbitmq-cluster-operator-779fc9694b-nzkln\" (UID: \"40c1982c-ddc0-4043-a956-2a5332612176\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.608422 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwlr\" (UniqueName: \"kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr\") pod \"rabbitmq-cluster-operator-779fc9694b-nzkln\" (UID: \"40c1982c-ddc0-4043-a956-2a5332612176\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:08:17 crc kubenswrapper[4967]: I0120 09:08:17.778868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.193722 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.246200 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.269876 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.397889 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" event={"ID":"40c1982c-ddc0-4043-a956-2a5332612176","Type":"ContainerStarted","Data":"af6937c7704c79a09486dfd57b27752ee614d0620e6d29903d8ec0e088402dcb"} Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.474320 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:08:18 crc kubenswrapper[4967]: I0120 09:08:18.474403 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:08:22 crc kubenswrapper[4967]: I0120 09:08:22.423264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" event={"ID":"40c1982c-ddc0-4043-a956-2a5332612176","Type":"ContainerStarted","Data":"6b4074506c651043bcbf53da34b1aa44d340376b57f491fe9a75004cd975fc1d"} Jan 20 09:08:22 crc kubenswrapper[4967]: I0120 09:08:22.440930 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" podStartSLOduration=2.046926269 podStartE2EDuration="5.440910816s" podCreationTimestamp="2026-01-20 09:08:17 +0000 UTC" firstStartedPulling="2026-01-20 09:08:18.258243856 +0000 UTC m=+1132.843544063" lastFinishedPulling="2026-01-20 09:08:21.652228403 +0000 UTC m=+1136.237528610" observedRunningTime="2026-01-20 09:08:22.435936888 +0000 UTC m=+1137.021237096" watchObservedRunningTime="2026-01-20 09:08:22.440910816 +0000 UTC m=+1137.026211023" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.077870 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.079215 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.084591 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"rabbitmq-plugins-conf" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.084822 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-erlang-cookie" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.084948 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-server-dockercfg-jhvbc" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.085063 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"rabbitmq-server-conf" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.085693 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"rabbitmq-default-user" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.094447 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168734 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168784 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168802 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjmmk\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168850 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168923 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168953 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.168972 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.270411 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271463 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271506 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271524 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271541 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjmmk\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271592 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271648 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.271669 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.272133 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.272913 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.273221 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.275552 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.275584 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c74ef7f3a5b7797511c2dea0330213a03ab988a7ccb6d042c7bee6c030df9c00/globalmount\"" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.294772 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.294867 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.295002 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.301657 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjmmk\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.320052 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") pod \"rabbitmq-server-0\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.403601 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:08:24 crc kubenswrapper[4967]: I0120 09:08:24.622181 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.443990 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerStarted","Data":"150862a26ff3a1ce99025897228e2d43625325b267eaf27152476d8d9f9465e8"} Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.717863 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.720416 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.722862 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-lh9x7" Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.725020 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.791191 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c64vr\" (UniqueName: \"kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr\") pod \"keystone-operator-index-wk787\" (UID: \"6042bd61-c186-4816-97a5-17fba8991f43\") " pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.892690 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c64vr\" (UniqueName: \"kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr\") pod \"keystone-operator-index-wk787\" (UID: \"6042bd61-c186-4816-97a5-17fba8991f43\") " pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:25 crc kubenswrapper[4967]: I0120 09:08:25.926338 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c64vr\" (UniqueName: \"kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr\") pod \"keystone-operator-index-wk787\" (UID: \"6042bd61-c186-4816-97a5-17fba8991f43\") " pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:26 crc kubenswrapper[4967]: I0120 09:08:26.047224 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:26 crc kubenswrapper[4967]: I0120 09:08:26.337424 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:26 crc kubenswrapper[4967]: W0120 09:08:26.372267 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6042bd61_c186_4816_97a5_17fba8991f43.slice/crio-f57e137c3736d7ab9a1d5d2eadd8116c74aa40ca3c1470c6baa12089393103ff WatchSource:0}: Error finding container f57e137c3736d7ab9a1d5d2eadd8116c74aa40ca3c1470c6baa12089393103ff: Status 404 returned error can't find the container with id f57e137c3736d7ab9a1d5d2eadd8116c74aa40ca3c1470c6baa12089393103ff Jan 20 09:08:26 crc kubenswrapper[4967]: I0120 09:08:26.450488 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-wk787" event={"ID":"6042bd61-c186-4816-97a5-17fba8991f43","Type":"ContainerStarted","Data":"f57e137c3736d7ab9a1d5d2eadd8116c74aa40ca3c1470c6baa12089393103ff"} Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.091486 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.699313 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.701933 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.705680 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.878818 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqf9m\" (UniqueName: \"kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m\") pod \"keystone-operator-index-kxddm\" (UID: \"01a3a285-14b5-4af7-87de-d3ef4c459794\") " pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:30 crc kubenswrapper[4967]: I0120 09:08:30.980711 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqf9m\" (UniqueName: \"kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m\") pod \"keystone-operator-index-kxddm\" (UID: \"01a3a285-14b5-4af7-87de-d3ef4c459794\") " pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:31 crc kubenswrapper[4967]: I0120 09:08:31.224073 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqf9m\" (UniqueName: \"kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m\") pod \"keystone-operator-index-kxddm\" (UID: \"01a3a285-14b5-4af7-87de-d3ef4c459794\") " pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:31 crc kubenswrapper[4967]: I0120 09:08:31.517033 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:35 crc kubenswrapper[4967]: I0120 09:08:35.961745 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:08:35 crc kubenswrapper[4967]: W0120 09:08:35.966361 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01a3a285_14b5_4af7_87de_d3ef4c459794.slice/crio-7cbd696484a3c1bb870ca33bb87b5aa78c0828013d7ae3f13c9c96bcd1b663ae WatchSource:0}: Error finding container 7cbd696484a3c1bb870ca33bb87b5aa78c0828013d7ae3f13c9c96bcd1b663ae: Status 404 returned error can't find the container with id 7cbd696484a3c1bb870ca33bb87b5aa78c0828013d7ae3f13c9c96bcd1b663ae Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.528671 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kxddm" event={"ID":"01a3a285-14b5-4af7-87de-d3ef4c459794","Type":"ContainerStarted","Data":"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f"} Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.528713 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kxddm" event={"ID":"01a3a285-14b5-4af7-87de-d3ef4c459794","Type":"ContainerStarted","Data":"7cbd696484a3c1bb870ca33bb87b5aa78c0828013d7ae3f13c9c96bcd1b663ae"} Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.530537 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-wk787" event={"ID":"6042bd61-c186-4816-97a5-17fba8991f43","Type":"ContainerStarted","Data":"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928"} Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.530640 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-wk787" podUID="6042bd61-c186-4816-97a5-17fba8991f43" containerName="registry-server" containerID="cri-o://fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928" gracePeriod=2 Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.550581 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-kxddm" podStartSLOduration=6.509860723 podStartE2EDuration="6.550565436s" podCreationTimestamp="2026-01-20 09:08:30 +0000 UTC" firstStartedPulling="2026-01-20 09:08:35.970657432 +0000 UTC m=+1150.555957639" lastFinishedPulling="2026-01-20 09:08:36.011362145 +0000 UTC m=+1150.596662352" observedRunningTime="2026-01-20 09:08:36.546175885 +0000 UTC m=+1151.131476082" watchObservedRunningTime="2026-01-20 09:08:36.550565436 +0000 UTC m=+1151.135865643" Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.566201 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-wk787" podStartSLOduration=2.419715873 podStartE2EDuration="11.566180857s" podCreationTimestamp="2026-01-20 09:08:25 +0000 UTC" firstStartedPulling="2026-01-20 09:08:26.37541102 +0000 UTC m=+1140.960711227" lastFinishedPulling="2026-01-20 09:08:35.521876004 +0000 UTC m=+1150.107176211" observedRunningTime="2026-01-20 09:08:36.561444636 +0000 UTC m=+1151.146744843" watchObservedRunningTime="2026-01-20 09:08:36.566180857 +0000 UTC m=+1151.151481064" Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.891452 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.966087 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c64vr\" (UniqueName: \"kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr\") pod \"6042bd61-c186-4816-97a5-17fba8991f43\" (UID: \"6042bd61-c186-4816-97a5-17fba8991f43\") " Jan 20 09:08:36 crc kubenswrapper[4967]: I0120 09:08:36.971353 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr" (OuterVolumeSpecName: "kube-api-access-c64vr") pod "6042bd61-c186-4816-97a5-17fba8991f43" (UID: "6042bd61-c186-4816-97a5-17fba8991f43"). InnerVolumeSpecName "kube-api-access-c64vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.067824 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c64vr\" (UniqueName: \"kubernetes.io/projected/6042bd61-c186-4816-97a5-17fba8991f43-kube-api-access-c64vr\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.537895 4967 generic.go:334] "Generic (PLEG): container finished" podID="6042bd61-c186-4816-97a5-17fba8991f43" containerID="fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928" exitCode=0 Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.537972 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-wk787" event={"ID":"6042bd61-c186-4816-97a5-17fba8991f43","Type":"ContainerDied","Data":"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928"} Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.538018 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-wk787" event={"ID":"6042bd61-c186-4816-97a5-17fba8991f43","Type":"ContainerDied","Data":"f57e137c3736d7ab9a1d5d2eadd8116c74aa40ca3c1470c6baa12089393103ff"} Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.538042 4967 scope.go:117] "RemoveContainer" containerID="fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.538178 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-wk787" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.539554 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerStarted","Data":"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90"} Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.557744 4967 scope.go:117] "RemoveContainer" containerID="fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928" Jan 20 09:08:37 crc kubenswrapper[4967]: E0120 09:08:37.559075 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928\": container with ID starting with fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928 not found: ID does not exist" containerID="fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.559139 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928"} err="failed to get container status \"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928\": rpc error: code = NotFound desc = could not find container \"fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928\": container with ID starting with fd7ef3f849916b60355d997a2ee7779d532ec4eaef86fda3c579fa9a11aef928 not found: ID does not exist" Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.581415 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.585671 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-wk787"] Jan 20 09:08:37 crc kubenswrapper[4967]: I0120 09:08:37.703063 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6042bd61-c186-4816-97a5-17fba8991f43" path="/var/lib/kubelet/pods/6042bd61-c186-4816-97a5-17fba8991f43/volumes" Jan 20 09:08:41 crc kubenswrapper[4967]: I0120 09:08:41.517713 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:41 crc kubenswrapper[4967]: I0120 09:08:41.518270 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:41 crc kubenswrapper[4967]: I0120 09:08:41.555733 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:41 crc kubenswrapper[4967]: I0120 09:08:41.600600 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.158804 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2"] Jan 20 09:08:45 crc kubenswrapper[4967]: E0120 09:08:45.159591 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6042bd61-c186-4816-97a5-17fba8991f43" containerName="registry-server" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.159624 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6042bd61-c186-4816-97a5-17fba8991f43" containerName="registry-server" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.159932 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6042bd61-c186-4816-97a5-17fba8991f43" containerName="registry-server" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.162586 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.167158 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lqxxp" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.174719 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2"] Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.280940 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.280991 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfxrm\" (UniqueName: \"kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.281033 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.382347 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.382396 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfxrm\" (UniqueName: \"kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.382439 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.382982 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.383046 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.405752 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfxrm\" (UniqueName: \"kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm\") pod \"62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.489739 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:45 crc kubenswrapper[4967]: I0120 09:08:45.926131 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2"] Jan 20 09:08:46 crc kubenswrapper[4967]: I0120 09:08:46.596462 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" event={"ID":"06f730d2-8988-4e09-9cb5-93670237e76a","Type":"ContainerStarted","Data":"1d4eb37fcbd8dcb9bf49afac1bee4dcf76e65dcc8a4c4902ea9b3d7c281a036e"} Jan 20 09:08:47 crc kubenswrapper[4967]: I0120 09:08:47.603200 4967 generic.go:334] "Generic (PLEG): container finished" podID="06f730d2-8988-4e09-9cb5-93670237e76a" containerID="69ded61430a547c32b4a3941a4ce7861ee7051fe4ad365a827603cfe78e49077" exitCode=0 Jan 20 09:08:47 crc kubenswrapper[4967]: I0120 09:08:47.603249 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" event={"ID":"06f730d2-8988-4e09-9cb5-93670237e76a","Type":"ContainerDied","Data":"69ded61430a547c32b4a3941a4ce7861ee7051fe4ad365a827603cfe78e49077"} Jan 20 09:08:47 crc kubenswrapper[4967]: I0120 09:08:47.604964 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 09:08:48 crc kubenswrapper[4967]: I0120 09:08:48.474423 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:08:48 crc kubenswrapper[4967]: I0120 09:08:48.474800 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:08:49 crc kubenswrapper[4967]: I0120 09:08:49.616288 4967 generic.go:334] "Generic (PLEG): container finished" podID="06f730d2-8988-4e09-9cb5-93670237e76a" containerID="f4b8eb2d6e37d1673728a01f88ac6deea3102110aad10b7a469da0c015bc751f" exitCode=0 Jan 20 09:08:49 crc kubenswrapper[4967]: I0120 09:08:49.616359 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" event={"ID":"06f730d2-8988-4e09-9cb5-93670237e76a","Type":"ContainerDied","Data":"f4b8eb2d6e37d1673728a01f88ac6deea3102110aad10b7a469da0c015bc751f"} Jan 20 09:08:50 crc kubenswrapper[4967]: I0120 09:08:50.625905 4967 generic.go:334] "Generic (PLEG): container finished" podID="06f730d2-8988-4e09-9cb5-93670237e76a" containerID="118445c68a573eb23fe42c671e11f042b7d6f627aa12473582d6379eb23eadc1" exitCode=0 Jan 20 09:08:50 crc kubenswrapper[4967]: I0120 09:08:50.625956 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" event={"ID":"06f730d2-8988-4e09-9cb5-93670237e76a","Type":"ContainerDied","Data":"118445c68a573eb23fe42c671e11f042b7d6f627aa12473582d6379eb23eadc1"} Jan 20 09:08:51 crc kubenswrapper[4967]: I0120 09:08:51.937295 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.069282 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfxrm\" (UniqueName: \"kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm\") pod \"06f730d2-8988-4e09-9cb5-93670237e76a\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.069392 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle\") pod \"06f730d2-8988-4e09-9cb5-93670237e76a\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.069498 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util\") pod \"06f730d2-8988-4e09-9cb5-93670237e76a\" (UID: \"06f730d2-8988-4e09-9cb5-93670237e76a\") " Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.070211 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle" (OuterVolumeSpecName: "bundle") pod "06f730d2-8988-4e09-9cb5-93670237e76a" (UID: "06f730d2-8988-4e09-9cb5-93670237e76a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.077812 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm" (OuterVolumeSpecName: "kube-api-access-tfxrm") pod "06f730d2-8988-4e09-9cb5-93670237e76a" (UID: "06f730d2-8988-4e09-9cb5-93670237e76a"). InnerVolumeSpecName "kube-api-access-tfxrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.083940 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util" (OuterVolumeSpecName: "util") pod "06f730d2-8988-4e09-9cb5-93670237e76a" (UID: "06f730d2-8988-4e09-9cb5-93670237e76a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.171226 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.171296 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/06f730d2-8988-4e09-9cb5-93670237e76a-util\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.171315 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfxrm\" (UniqueName: \"kubernetes.io/projected/06f730d2-8988-4e09-9cb5-93670237e76a-kube-api-access-tfxrm\") on node \"crc\" DevicePath \"\"" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.640781 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" event={"ID":"06f730d2-8988-4e09-9cb5-93670237e76a","Type":"ContainerDied","Data":"1d4eb37fcbd8dcb9bf49afac1bee4dcf76e65dcc8a4c4902ea9b3d7c281a036e"} Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.640824 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d4eb37fcbd8dcb9bf49afac1bee4dcf76e65dcc8a4c4902ea9b3d7c281a036e" Jan 20 09:08:52 crc kubenswrapper[4967]: I0120 09:08:52.640852 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.950456 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:09:01 crc kubenswrapper[4967]: E0120 09:09:01.951341 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="pull" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.951354 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="pull" Jan 20 09:09:01 crc kubenswrapper[4967]: E0120 09:09:01.951368 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="util" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.951376 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="util" Jan 20 09:09:01 crc kubenswrapper[4967]: E0120 09:09:01.951400 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="extract" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.951419 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="extract" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.951675 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" containerName="extract" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.952296 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.959733 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-d9tpg" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.963624 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Jan 20 09:09:01 crc kubenswrapper[4967]: I0120 09:09:01.976459 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.102248 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6bpp\" (UniqueName: \"kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.102331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.102368 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.203237 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.203335 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6bpp\" (UniqueName: \"kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.203395 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.209233 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.209564 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.219644 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6bpp\" (UniqueName: \"kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp\") pod \"keystone-operator-controller-manager-74884b4b57-zjgsp\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.281628 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.690348 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:09:02 crc kubenswrapper[4967]: I0120 09:09:02.698336 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" event={"ID":"b7bbfe3d-10a4-497f-87e0-d35c154af3ca","Type":"ContainerStarted","Data":"199e8252c9af244d80edbb1e24e424f1406aee7a9c079290b09f44916e2a2900"} Jan 20 09:09:06 crc kubenswrapper[4967]: I0120 09:09:06.744479 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" event={"ID":"b7bbfe3d-10a4-497f-87e0-d35c154af3ca","Type":"ContainerStarted","Data":"064ee451f731c19f486eac61edbcd28e00e7375cb7bc819851f112a3150b736d"} Jan 20 09:09:06 crc kubenswrapper[4967]: I0120 09:09:06.755853 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:06 crc kubenswrapper[4967]: I0120 09:09:06.778085 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" podStartSLOduration=2.2864144619999998 podStartE2EDuration="5.778071103s" podCreationTimestamp="2026-01-20 09:09:01 +0000 UTC" firstStartedPulling="2026-01-20 09:09:02.689131064 +0000 UTC m=+1177.274431271" lastFinishedPulling="2026-01-20 09:09:06.180787705 +0000 UTC m=+1180.766087912" observedRunningTime="2026-01-20 09:09:06.777402403 +0000 UTC m=+1181.362702610" watchObservedRunningTime="2026-01-20 09:09:06.778071103 +0000 UTC m=+1181.363371310" Jan 20 09:09:08 crc kubenswrapper[4967]: I0120 09:09:08.757200 4967 generic.go:334] "Generic (PLEG): container finished" podID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerID="2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90" exitCode=0 Jan 20 09:09:08 crc kubenswrapper[4967]: I0120 09:09:08.757288 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerDied","Data":"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90"} Jan 20 09:09:09 crc kubenswrapper[4967]: I0120 09:09:09.766514 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerStarted","Data":"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715"} Jan 20 09:09:09 crc kubenswrapper[4967]: I0120 09:09:09.767669 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:09:09 crc kubenswrapper[4967]: I0120 09:09:09.791696 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/rabbitmq-server-0" podStartSLOduration=35.514874381 podStartE2EDuration="46.79167659s" podCreationTimestamp="2026-01-20 09:08:23 +0000 UTC" firstStartedPulling="2026-01-20 09:08:24.630859055 +0000 UTC m=+1139.216159262" lastFinishedPulling="2026-01-20 09:08:35.907661264 +0000 UTC m=+1150.492961471" observedRunningTime="2026-01-20 09:09:09.78981967 +0000 UTC m=+1184.375119927" watchObservedRunningTime="2026-01-20 09:09:09.79167659 +0000 UTC m=+1184.376976807" Jan 20 09:09:12 crc kubenswrapper[4967]: I0120 09:09:12.287162 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.910485 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-2x79z"] Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.912022 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.915954 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2"] Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.917041 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.921137 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-2x79z"] Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.921934 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Jan 20 09:09:17 crc kubenswrapper[4967]: I0120 09:09:17.927736 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2"] Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.004738 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmfdc\" (UniqueName: \"kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.004820 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.004852 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjkm\" (UniqueName: \"kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.004950 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.106283 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.106340 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjkm\" (UniqueName: \"kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.106421 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.106462 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmfdc\" (UniqueName: \"kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.107166 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.107310 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.126834 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmfdc\" (UniqueName: \"kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc\") pod \"keystone-a661-account-create-update-k9zt2\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.131434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjkm\" (UniqueName: \"kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm\") pod \"keystone-db-create-2x79z\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.246984 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.254515 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.476892 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.477218 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.477260 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.477868 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.477916 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3" gracePeriod=600 Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.834966 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-2x79z"] Jan 20 09:09:18 crc kubenswrapper[4967]: I0120 09:09:18.888623 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2"] Jan 20 09:09:18 crc kubenswrapper[4967]: W0120 09:09:18.892256 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf5b5257_3852_4cd8_8085_558c8e9bd5d5.slice/crio-5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb WatchSource:0}: Error finding container 5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb: Status 404 returned error can't find the container with id 5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.832368 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3" exitCode=0 Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.832461 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3"} Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.833425 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e"} Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.833445 4967 scope.go:117] "RemoveContainer" containerID="c9e6dee324dc4b0a5a95aadbf0233c2780515142335da9d7011334887dda237e" Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.836198 4967 generic.go:334] "Generic (PLEG): container finished" podID="0e488abc-d383-4d59-a608-aeeb30111be3" containerID="41b97d5eba37826f4dbc0832c212b967ba13596c23b3a4d9d292f447ca92f549" exitCode=0 Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.836233 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-2x79z" event={"ID":"0e488abc-d383-4d59-a608-aeeb30111be3","Type":"ContainerDied","Data":"41b97d5eba37826f4dbc0832c212b967ba13596c23b3a4d9d292f447ca92f549"} Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.836276 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-2x79z" event={"ID":"0e488abc-d383-4d59-a608-aeeb30111be3","Type":"ContainerStarted","Data":"df92938f72a72dd4250363481675286fedde8ba553d72c3d93afa758012e372d"} Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.845282 4967 generic.go:334] "Generic (PLEG): container finished" podID="cf5b5257-3852-4cd8-8085-558c8e9bd5d5" containerID="9a8d317eeb3fad645fc67e686323a9316759fa2785ce437b9e90e9f5f30854a9" exitCode=0 Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.845333 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" event={"ID":"cf5b5257-3852-4cd8-8085-558c8e9bd5d5","Type":"ContainerDied","Data":"9a8d317eeb3fad645fc67e686323a9316759fa2785ce437b9e90e9f5f30854a9"} Jan 20 09:09:19 crc kubenswrapper[4967]: I0120 09:09:19.845381 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" event={"ID":"cf5b5257-3852-4cd8-8085-558c8e9bd5d5","Type":"ContainerStarted","Data":"5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb"} Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.192869 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.249596 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmfdc\" (UniqueName: \"kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc\") pod \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.250018 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts\") pod \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\" (UID: \"cf5b5257-3852-4cd8-8085-558c8e9bd5d5\") " Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.250793 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf5b5257-3852-4cd8-8085-558c8e9bd5d5" (UID: "cf5b5257-3852-4cd8-8085-558c8e9bd5d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.255872 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc" (OuterVolumeSpecName: "kube-api-access-pmfdc") pod "cf5b5257-3852-4cd8-8085-558c8e9bd5d5" (UID: "cf5b5257-3852-4cd8-8085-558c8e9bd5d5"). InnerVolumeSpecName "kube-api-access-pmfdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.286963 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.352635 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts\") pod \"0e488abc-d383-4d59-a608-aeeb30111be3\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.352756 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mjkm\" (UniqueName: \"kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm\") pod \"0e488abc-d383-4d59-a608-aeeb30111be3\" (UID: \"0e488abc-d383-4d59-a608-aeeb30111be3\") " Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.353317 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmfdc\" (UniqueName: \"kubernetes.io/projected/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-kube-api-access-pmfdc\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.353344 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf5b5257-3852-4cd8-8085-558c8e9bd5d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.353570 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e488abc-d383-4d59-a608-aeeb30111be3" (UID: "0e488abc-d383-4d59-a608-aeeb30111be3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.356524 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm" (OuterVolumeSpecName: "kube-api-access-4mjkm") pod "0e488abc-d383-4d59-a608-aeeb30111be3" (UID: "0e488abc-d383-4d59-a608-aeeb30111be3"). InnerVolumeSpecName "kube-api-access-4mjkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.455249 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e488abc-d383-4d59-a608-aeeb30111be3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.455318 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mjkm\" (UniqueName: \"kubernetes.io/projected/0e488abc-d383-4d59-a608-aeeb30111be3-kube-api-access-4mjkm\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.862205 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" event={"ID":"cf5b5257-3852-4cd8-8085-558c8e9bd5d5","Type":"ContainerDied","Data":"5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb"} Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.862594 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cf04e200f3b7e34af649e69d233fe591d350f223d3735a68bb74e587c64b3eb" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.862366 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.864331 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-2x79z" event={"ID":"0e488abc-d383-4d59-a608-aeeb30111be3","Type":"ContainerDied","Data":"df92938f72a72dd4250363481675286fedde8ba553d72c3d93afa758012e372d"} Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.864382 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df92938f72a72dd4250363481675286fedde8ba553d72c3d93afa758012e372d" Jan 20 09:09:21 crc kubenswrapper[4967]: I0120 09:09:21.864435 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-2x79z" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.406455 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.939188 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-ltncp"] Jan 20 09:09:24 crc kubenswrapper[4967]: E0120 09:09:24.939712 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5b5257-3852-4cd8-8085-558c8e9bd5d5" containerName="mariadb-account-create-update" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.939730 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5b5257-3852-4cd8-8085-558c8e9bd5d5" containerName="mariadb-account-create-update" Jan 20 09:09:24 crc kubenswrapper[4967]: E0120 09:09:24.939752 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e488abc-d383-4d59-a608-aeeb30111be3" containerName="mariadb-database-create" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.939759 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e488abc-d383-4d59-a608-aeeb30111be3" containerName="mariadb-database-create" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.939864 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e488abc-d383-4d59-a608-aeeb30111be3" containerName="mariadb-database-create" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.939878 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5b5257-3852-4cd8-8085-558c8e9bd5d5" containerName="mariadb-account-create-update" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.940279 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.944043 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.944243 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.944662 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-548r8" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.944808 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:09:24 crc kubenswrapper[4967]: I0120 09:09:24.948924 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-ltncp"] Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.007875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc9sd\" (UniqueName: \"kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.008063 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.109842 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.109924 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc9sd\" (UniqueName: \"kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.115592 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.126859 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc9sd\" (UniqueName: \"kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd\") pod \"keystone-db-sync-ltncp\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.264698 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.652852 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-ltncp"] Jan 20 09:09:25 crc kubenswrapper[4967]: I0120 09:09:25.887747 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" event={"ID":"20c308be-da30-474a-9446-aee0ea84af9d","Type":"ContainerStarted","Data":"ebfd8b4ca4c27deaed20b8656ef715c383cb067e235021e5713fa1c474d6fb39"} Jan 20 09:09:34 crc kubenswrapper[4967]: I0120 09:09:34.953652 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" event={"ID":"20c308be-da30-474a-9446-aee0ea84af9d","Type":"ContainerStarted","Data":"911d966924f4b1c670c0c0ca4760ce4e647853a40d3f2c6f0e4f12830e38dcdf"} Jan 20 09:09:37 crc kubenswrapper[4967]: I0120 09:09:37.972422 4967 generic.go:334] "Generic (PLEG): container finished" podID="20c308be-da30-474a-9446-aee0ea84af9d" containerID="911d966924f4b1c670c0c0ca4760ce4e647853a40d3f2c6f0e4f12830e38dcdf" exitCode=0 Jan 20 09:09:37 crc kubenswrapper[4967]: I0120 09:09:37.972516 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" event={"ID":"20c308be-da30-474a-9446-aee0ea84af9d","Type":"ContainerDied","Data":"911d966924f4b1c670c0c0ca4760ce4e647853a40d3f2c6f0e4f12830e38dcdf"} Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.240819 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.434845 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data\") pod \"20c308be-da30-474a-9446-aee0ea84af9d\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.435000 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc9sd\" (UniqueName: \"kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd\") pod \"20c308be-da30-474a-9446-aee0ea84af9d\" (UID: \"20c308be-da30-474a-9446-aee0ea84af9d\") " Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.443046 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd" (OuterVolumeSpecName: "kube-api-access-mc9sd") pod "20c308be-da30-474a-9446-aee0ea84af9d" (UID: "20c308be-da30-474a-9446-aee0ea84af9d"). InnerVolumeSpecName "kube-api-access-mc9sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.477768 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data" (OuterVolumeSpecName: "config-data") pod "20c308be-da30-474a-9446-aee0ea84af9d" (UID: "20c308be-da30-474a-9446-aee0ea84af9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.536935 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c308be-da30-474a-9446-aee0ea84af9d-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.536964 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc9sd\" (UniqueName: \"kubernetes.io/projected/20c308be-da30-474a-9446-aee0ea84af9d-kube-api-access-mc9sd\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.985040 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" event={"ID":"20c308be-da30-474a-9446-aee0ea84af9d","Type":"ContainerDied","Data":"ebfd8b4ca4c27deaed20b8656ef715c383cb067e235021e5713fa1c474d6fb39"} Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.985066 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-ltncp" Jan 20 09:09:39 crc kubenswrapper[4967]: I0120 09:09:39.985082 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebfd8b4ca4c27deaed20b8656ef715c383cb067e235021e5713fa1c474d6fb39" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.168348 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-pr8nf"] Jan 20 09:09:40 crc kubenswrapper[4967]: E0120 09:09:40.168668 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c308be-da30-474a-9446-aee0ea84af9d" containerName="keystone-db-sync" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.168686 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c308be-da30-474a-9446-aee0ea84af9d" containerName="keystone-db-sync" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.168833 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c308be-da30-474a-9446-aee0ea84af9d" containerName="keystone-db-sync" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.169365 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.172199 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-548r8" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.172489 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.172675 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.175672 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.175796 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-pr8nf"] Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.177905 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.246898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.246946 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.246982 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.247014 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lcvs\" (UniqueName: \"kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.247070 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.347489 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lcvs\" (UniqueName: \"kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.347782 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.347951 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.348059 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.348173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.359380 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.359500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.359512 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.359562 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.364513 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lcvs\" (UniqueName: \"kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs\") pod \"keystone-bootstrap-pr8nf\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:40 crc kubenswrapper[4967]: I0120 09:09:40.484543 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:41 crc kubenswrapper[4967]: I0120 09:09:41.122833 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-pr8nf"] Jan 20 09:09:41 crc kubenswrapper[4967]: I0120 09:09:41.998468 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" event={"ID":"9551bc9a-08bc-49d1-b567-a4e6593c49f0","Type":"ContainerStarted","Data":"d0f8575ff69255bd79a19bdb792557eebfc4e793602739a31da09252c42c0062"} Jan 20 09:09:41 crc kubenswrapper[4967]: I0120 09:09:41.998782 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" event={"ID":"9551bc9a-08bc-49d1-b567-a4e6593c49f0","Type":"ContainerStarted","Data":"08c494f29d2e0a3c8adedfc00b87be084ee681700a6f4e1326b5a5093bf04cbe"} Jan 20 09:09:42 crc kubenswrapper[4967]: I0120 09:09:42.011682 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" podStartSLOduration=2.01166109 podStartE2EDuration="2.01166109s" podCreationTimestamp="2026-01-20 09:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:09:42.011414223 +0000 UTC m=+1216.596714430" watchObservedRunningTime="2026-01-20 09:09:42.01166109 +0000 UTC m=+1216.596961297" Jan 20 09:09:44 crc kubenswrapper[4967]: I0120 09:09:44.010171 4967 generic.go:334] "Generic (PLEG): container finished" podID="9551bc9a-08bc-49d1-b567-a4e6593c49f0" containerID="d0f8575ff69255bd79a19bdb792557eebfc4e793602739a31da09252c42c0062" exitCode=0 Jan 20 09:09:44 crc kubenswrapper[4967]: I0120 09:09:44.010266 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" event={"ID":"9551bc9a-08bc-49d1-b567-a4e6593c49f0","Type":"ContainerDied","Data":"d0f8575ff69255bd79a19bdb792557eebfc4e793602739a31da09252c42c0062"} Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.289599 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.423857 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys\") pod \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.424004 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts\") pod \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.424144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys\") pod \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.424184 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data\") pod \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.424266 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lcvs\" (UniqueName: \"kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs\") pod \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\" (UID: \"9551bc9a-08bc-49d1-b567-a4e6593c49f0\") " Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.430344 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9551bc9a-08bc-49d1-b567-a4e6593c49f0" (UID: "9551bc9a-08bc-49d1-b567-a4e6593c49f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.431037 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts" (OuterVolumeSpecName: "scripts") pod "9551bc9a-08bc-49d1-b567-a4e6593c49f0" (UID: "9551bc9a-08bc-49d1-b567-a4e6593c49f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.432237 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9551bc9a-08bc-49d1-b567-a4e6593c49f0" (UID: "9551bc9a-08bc-49d1-b567-a4e6593c49f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.433874 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs" (OuterVolumeSpecName: "kube-api-access-9lcvs") pod "9551bc9a-08bc-49d1-b567-a4e6593c49f0" (UID: "9551bc9a-08bc-49d1-b567-a4e6593c49f0"). InnerVolumeSpecName "kube-api-access-9lcvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.449018 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data" (OuterVolumeSpecName: "config-data") pod "9551bc9a-08bc-49d1-b567-a4e6593c49f0" (UID: "9551bc9a-08bc-49d1-b567-a4e6593c49f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.526652 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.526684 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.526693 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.526703 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9551bc9a-08bc-49d1-b567-a4e6593c49f0-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:45 crc kubenswrapper[4967]: I0120 09:09:45.526715 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lcvs\" (UniqueName: \"kubernetes.io/projected/9551bc9a-08bc-49d1-b567-a4e6593c49f0-kube-api-access-9lcvs\") on node \"crc\" DevicePath \"\"" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.027025 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" event={"ID":"9551bc9a-08bc-49d1-b567-a4e6593c49f0","Type":"ContainerDied","Data":"08c494f29d2e0a3c8adedfc00b87be084ee681700a6f4e1326b5a5093bf04cbe"} Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.027347 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08c494f29d2e0a3c8adedfc00b87be084ee681700a6f4e1326b5a5093bf04cbe" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.027315 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-pr8nf" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.109824 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:09:46 crc kubenswrapper[4967]: E0120 09:09:46.110062 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9551bc9a-08bc-49d1-b567-a4e6593c49f0" containerName="keystone-bootstrap" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.110077 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9551bc9a-08bc-49d1-b567-a4e6593c49f0" containerName="keystone-bootstrap" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.110201 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9551bc9a-08bc-49d1-b567-a4e6593c49f0" containerName="keystone-bootstrap" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.110637 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.112186 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-548r8" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.117140 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.141275 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.141677 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.142390 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.233846 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzrvl\" (UniqueName: \"kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.233891 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.233936 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.233955 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.233974 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.338309 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.338390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.338410 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.338427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.338667 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzrvl\" (UniqueName: \"kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.347681 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.349352 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.349863 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.350458 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.356587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzrvl\" (UniqueName: \"kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl\") pod \"keystone-74cf68b967-vqrkn\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.460334 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:46 crc kubenswrapper[4967]: I0120 09:09:46.658676 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:09:47 crc kubenswrapper[4967]: I0120 09:09:47.035548 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" event={"ID":"2b7f857e-08ef-4366-b82c-345bea975113","Type":"ContainerStarted","Data":"24fc2b9babcb2a3bc2223f55558ff65d5fcaea01d84fbdd4c1732a63548b8542"} Jan 20 09:09:47 crc kubenswrapper[4967]: I0120 09:09:47.035922 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:09:47 crc kubenswrapper[4967]: I0120 09:09:47.035945 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" event={"ID":"2b7f857e-08ef-4366-b82c-345bea975113","Type":"ContainerStarted","Data":"8cc3b75d3e809c4a707f15238947cb315edaffb5f4c1058f956de327339c6a35"} Jan 20 09:09:47 crc kubenswrapper[4967]: I0120 09:09:47.058400 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" podStartSLOduration=1.05837733 podStartE2EDuration="1.05837733s" podCreationTimestamp="2026-01-20 09:09:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:09:47.05473992 +0000 UTC m=+1221.640040167" watchObservedRunningTime="2026-01-20 09:09:47.05837733 +0000 UTC m=+1221.643677567" Jan 20 09:10:17 crc kubenswrapper[4967]: I0120 09:10:17.950138 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:10:19 crc kubenswrapper[4967]: E0120 09:10:19.018685 4967 log.go:32] "Failed when writing line to log file" err="http2: stream closed" path="/var/log/pods/keystone-kuttl-tests_keystone-74cf68b967-vqrkn_2b7f857e-08ef-4366-b82c-345bea975113/keystone-api/0.log" line={} Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.359898 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.360628 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.381665 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.381710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkssw\" (UniqueName: \"kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.381733 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.381773 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.381811 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.385068 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:19 crc kubenswrapper[4967]: E0120 09:10:19.452731 4967 log.go:32] "Failed when writing line to log file" err="http2: stream closed" path="/var/log/pods/keystone-kuttl-tests_keystone-74cf68b967-vqrkn_2b7f857e-08ef-4366-b82c-345bea975113/keystone-api/0.log" line={} Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.483602 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.483652 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkssw\" (UniqueName: \"kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.483674 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.483720 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.483743 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.488891 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.488955 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.489399 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.494733 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.501670 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkssw\" (UniqueName: \"kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw\") pod \"keystone-575ffd58d4-tj5dg\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:19 crc kubenswrapper[4967]: I0120 09:10:19.723499 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.104310 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.265829 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" event={"ID":"72511dbd-0c25-40b8-bbd8-786edeb23c1d","Type":"ContainerStarted","Data":"9c38e8c37cd655d3982686f7e11facb81a61f8d2262c93ecb9ad7867189f49b5"} Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.853906 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-ltncp"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.865896 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-ltncp"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.873987 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-pr8nf"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.883621 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-pr8nf"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.897926 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.898198 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" podUID="2b7f857e-08ef-4366-b82c-345bea975113" containerName="keystone-api" containerID="cri-o://24fc2b9babcb2a3bc2223f55558ff65d5fcaea01d84fbdd4c1732a63548b8542" gracePeriod=30 Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.914361 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.919192 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystonea661-account-delete-zw27v"] Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.920896 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:20 crc kubenswrapper[4967]: I0120 09:10:20.923588 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonea661-account-delete-zw27v"] Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.001602 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.001708 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqtwp\" (UniqueName: \"kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.103032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqtwp\" (UniqueName: \"kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.103432 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.104283 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.126380 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqtwp\" (UniqueName: \"kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp\") pod \"keystonea661-account-delete-zw27v\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.250064 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.274210 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" event={"ID":"72511dbd-0c25-40b8-bbd8-786edeb23c1d","Type":"ContainerStarted","Data":"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769"} Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.274336 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.274553 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" secret="" err="secret \"keystone-keystone-dockercfg-548r8\" not found" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.288987 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" podStartSLOduration=2.28896743 podStartE2EDuration="2.28896743s" podCreationTimestamp="2026-01-20 09:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:10:21.286932615 +0000 UTC m=+1255.872232872" watchObservedRunningTime="2026-01-20 09:10:21.28896743 +0000 UTC m=+1255.874267647" Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308194 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308268 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:21.80824838 +0000 UTC m=+1256.393548587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308504 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-config-data: secret "keystone-config-data" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308517 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308547 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:21.808538187 +0000 UTC m=+1256.393838404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308565 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:21.808557768 +0000 UTC m=+1256.393858095 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone-config-data" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308639 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-scripts: secret "keystone-scripts" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.308710 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:21.808699162 +0000 UTC m=+1256.393999369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone-scripts" not found Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.453191 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonea661-account-delete-zw27v"] Jan 20 09:10:21 crc kubenswrapper[4967]: W0120 09:10:21.456534 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d3762f2_9c1c_40ef_b92b_e7a57a07bf81.slice/crio-92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4 WatchSource:0}: Error finding container 92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4: Status 404 returned error can't find the container with id 92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4 Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.702005 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c308be-da30-474a-9446-aee0ea84af9d" path="/var/lib/kubelet/pods/20c308be-da30-474a-9446-aee0ea84af9d/volumes" Jan 20 09:10:21 crc kubenswrapper[4967]: I0120 09:10:21.702709 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9551bc9a-08bc-49d1-b567-a4e6593c49f0" path="/var/lib/kubelet/pods/9551bc9a-08bc-49d1-b567-a4e6593c49f0/volumes" Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814398 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814458 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:22.814444556 +0000 UTC m=+1257.399744763 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "credential-keys" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814760 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone: secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814785 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:22.814778555 +0000 UTC m=+1257.400078762 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "fernet-keys" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814398 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-config-data: secret "keystone-config-data" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814807 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:22.814802065 +0000 UTC m=+1257.400102272 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone-config-data" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.814964 4967 secret.go:188] Couldn't get secret keystone-kuttl-tests/keystone-scripts: secret "keystone-scripts" not found Jan 20 09:10:21 crc kubenswrapper[4967]: E0120 09:10:21.815075 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts podName:72511dbd-0c25-40b8-bbd8-786edeb23c1d nodeName:}" failed. No retries permitted until 2026-01-20 09:10:22.815062392 +0000 UTC m=+1257.400362599 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts") pod "keystone-575ffd58d4-tj5dg" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d") : secret "keystone-scripts" not found Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.283661 4967 generic.go:334] "Generic (PLEG): container finished" podID="2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" containerID="bc81d8e4fb4fae0f2bec4e802abc804c4b1f66342f49d35ae19abcc2704f9783" exitCode=0 Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.284885 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" podUID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" containerName="keystone-api" containerID="cri-o://cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769" gracePeriod=30 Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.285194 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" event={"ID":"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81","Type":"ContainerDied","Data":"bc81d8e4fb4fae0f2bec4e802abc804c4b1f66342f49d35ae19abcc2704f9783"} Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.285223 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" event={"ID":"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81","Type":"ContainerStarted","Data":"92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4"} Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.625340 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.828027 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkssw\" (UniqueName: \"kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw\") pod \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.828154 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts\") pod \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.829838 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys\") pod \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.829928 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys\") pod \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.830031 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data\") pod \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\" (UID: \"72511dbd-0c25-40b8-bbd8-786edeb23c1d\") " Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.835141 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "72511dbd-0c25-40b8-bbd8-786edeb23c1d" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.836724 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "72511dbd-0c25-40b8-bbd8-786edeb23c1d" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.842744 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts" (OuterVolumeSpecName: "scripts") pod "72511dbd-0c25-40b8-bbd8-786edeb23c1d" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.842931 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw" (OuterVolumeSpecName: "kube-api-access-hkssw") pod "72511dbd-0c25-40b8-bbd8-786edeb23c1d" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d"). InnerVolumeSpecName "kube-api-access-hkssw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.855264 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data" (OuterVolumeSpecName: "config-data") pod "72511dbd-0c25-40b8-bbd8-786edeb23c1d" (UID: "72511dbd-0c25-40b8-bbd8-786edeb23c1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.932604 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkssw\" (UniqueName: \"kubernetes.io/projected/72511dbd-0c25-40b8-bbd8-786edeb23c1d-kube-api-access-hkssw\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.932662 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.932676 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.932690 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:22 crc kubenswrapper[4967]: I0120 09:10:22.932703 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72511dbd-0c25-40b8-bbd8-786edeb23c1d-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.292170 4967 generic.go:334] "Generic (PLEG): container finished" podID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" containerID="cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769" exitCode=0 Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.292221 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" event={"ID":"72511dbd-0c25-40b8-bbd8-786edeb23c1d","Type":"ContainerDied","Data":"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769"} Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.292235 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.293423 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-575ffd58d4-tj5dg" event={"ID":"72511dbd-0c25-40b8-bbd8-786edeb23c1d","Type":"ContainerDied","Data":"9c38e8c37cd655d3982686f7e11facb81a61f8d2262c93ecb9ad7867189f49b5"} Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.293442 4967 scope.go:117] "RemoveContainer" containerID="cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.319718 4967 scope.go:117] "RemoveContainer" containerID="cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769" Jan 20 09:10:23 crc kubenswrapper[4967]: E0120 09:10:23.320225 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769\": container with ID starting with cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769 not found: ID does not exist" containerID="cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.320269 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769"} err="failed to get container status \"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769\": rpc error: code = NotFound desc = could not find container \"cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769\": container with ID starting with cd35800d0a435367251906856371d3a8e42e4acf894e70a346f70229495e3769 not found: ID does not exist" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.325331 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.331650 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-575ffd58d4-tj5dg"] Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.521569 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.540595 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts\") pod \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.541299 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" (UID: "2d3762f2-9c1c-40ef-b92b-e7a57a07bf81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.642302 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqtwp\" (UniqueName: \"kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp\") pod \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\" (UID: \"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81\") " Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.642901 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.646789 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp" (OuterVolumeSpecName: "kube-api-access-hqtwp") pod "2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" (UID: "2d3762f2-9c1c-40ef-b92b-e7a57a07bf81"). InnerVolumeSpecName "kube-api-access-hqtwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.706172 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" path="/var/lib/kubelet/pods/72511dbd-0c25-40b8-bbd8-786edeb23c1d/volumes" Jan 20 09:10:23 crc kubenswrapper[4967]: I0120 09:10:23.744734 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqtwp\" (UniqueName: \"kubernetes.io/projected/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81-kube-api-access-hqtwp\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:24 crc kubenswrapper[4967]: I0120 09:10:24.303264 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" Jan 20 09:10:24 crc kubenswrapper[4967]: I0120 09:10:24.303311 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonea661-account-delete-zw27v" event={"ID":"2d3762f2-9c1c-40ef-b92b-e7a57a07bf81","Type":"ContainerDied","Data":"92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4"} Jan 20 09:10:24 crc kubenswrapper[4967]: I0120 09:10:24.303686 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92c3e30dc5fffcc07345c944e9aa9b00668e5caf8b7b26926512f3af4a76d0f4" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.311259 4967 generic.go:334] "Generic (PLEG): container finished" podID="2b7f857e-08ef-4366-b82c-345bea975113" containerID="24fc2b9babcb2a3bc2223f55558ff65d5fcaea01d84fbdd4c1732a63548b8542" exitCode=0 Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.311292 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" event={"ID":"2b7f857e-08ef-4366-b82c-345bea975113","Type":"ContainerDied","Data":"24fc2b9babcb2a3bc2223f55558ff65d5fcaea01d84fbdd4c1732a63548b8542"} Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.377153 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.469966 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts\") pod \"2b7f857e-08ef-4366-b82c-345bea975113\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.470028 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data\") pod \"2b7f857e-08ef-4366-b82c-345bea975113\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.470049 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzrvl\" (UniqueName: \"kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl\") pod \"2b7f857e-08ef-4366-b82c-345bea975113\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.470072 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys\") pod \"2b7f857e-08ef-4366-b82c-345bea975113\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.470090 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys\") pod \"2b7f857e-08ef-4366-b82c-345bea975113\" (UID: \"2b7f857e-08ef-4366-b82c-345bea975113\") " Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.474685 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2b7f857e-08ef-4366-b82c-345bea975113" (UID: "2b7f857e-08ef-4366-b82c-345bea975113"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.474738 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl" (OuterVolumeSpecName: "kube-api-access-vzrvl") pod "2b7f857e-08ef-4366-b82c-345bea975113" (UID: "2b7f857e-08ef-4366-b82c-345bea975113"). InnerVolumeSpecName "kube-api-access-vzrvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.475729 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts" (OuterVolumeSpecName: "scripts") pod "2b7f857e-08ef-4366-b82c-345bea975113" (UID: "2b7f857e-08ef-4366-b82c-345bea975113"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.476416 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2b7f857e-08ef-4366-b82c-345bea975113" (UID: "2b7f857e-08ef-4366-b82c-345bea975113"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.487358 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data" (OuterVolumeSpecName: "config-data") pod "2b7f857e-08ef-4366-b82c-345bea975113" (UID: "2b7f857e-08ef-4366-b82c-345bea975113"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.571189 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.571238 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzrvl\" (UniqueName: \"kubernetes.io/projected/2b7f857e-08ef-4366-b82c-345bea975113-kube-api-access-vzrvl\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.571251 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.571271 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.571282 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b7f857e-08ef-4366-b82c-345bea975113-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.907189 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-2x79z"] Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.911632 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-2x79z"] Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.920491 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2"] Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.925432 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystonea661-account-delete-zw27v"] Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.929768 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-a661-account-create-update-k9zt2"] Jan 20 09:10:25 crc kubenswrapper[4967]: I0120 09:10:25.933259 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystonea661-account-delete-zw27v"] Jan 20 09:10:26 crc kubenswrapper[4967]: I0120 09:10:26.318102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" event={"ID":"2b7f857e-08ef-4366-b82c-345bea975113","Type":"ContainerDied","Data":"8cc3b75d3e809c4a707f15238947cb315edaffb5f4c1058f956de327339c6a35"} Jan 20 09:10:26 crc kubenswrapper[4967]: I0120 09:10:26.318156 4967 scope.go:117] "RemoveContainer" containerID="24fc2b9babcb2a3bc2223f55558ff65d5fcaea01d84fbdd4c1732a63548b8542" Jan 20 09:10:26 crc kubenswrapper[4967]: I0120 09:10:26.318269 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-74cf68b967-vqrkn" Jan 20 09:10:26 crc kubenswrapper[4967]: I0120 09:10:26.344323 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:10:26 crc kubenswrapper[4967]: I0120 09:10:26.350200 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-74cf68b967-vqrkn"] Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184347 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ml9dr"] Jan 20 09:10:27 crc kubenswrapper[4967]: E0120 09:10:27.184681 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184703 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: E0120 09:10:27.184720 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" containerName="mariadb-account-delete" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184729 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" containerName="mariadb-account-delete" Jan 20 09:10:27 crc kubenswrapper[4967]: E0120 09:10:27.184748 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7f857e-08ef-4366-b82c-345bea975113" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184756 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7f857e-08ef-4366-b82c-345bea975113" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184943 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" containerName="mariadb-account-delete" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184960 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7f857e-08ef-4366-b82c-345bea975113" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.184969 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="72511dbd-0c25-40b8-bbd8-786edeb23c1d" containerName="keystone-api" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.185473 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.191313 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.191462 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-252ql\" (UniqueName: \"kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.193973 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ml9dr"] Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.198451 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-4941-account-create-update-7s89w"] Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.199246 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.201607 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.223390 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-4941-account-create-update-7s89w"] Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.292229 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-252ql\" (UniqueName: \"kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.292486 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xkdf\" (UniqueName: \"kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.292521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.292579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.293204 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.310690 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-252ql\" (UniqueName: \"kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql\") pod \"keystone-db-create-ml9dr\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.393482 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.393632 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xkdf\" (UniqueName: \"kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.394284 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.411582 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xkdf\" (UniqueName: \"kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf\") pod \"keystone-4941-account-create-update-7s89w\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.506903 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.537688 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.710871 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e488abc-d383-4d59-a608-aeeb30111be3" path="/var/lib/kubelet/pods/0e488abc-d383-4d59-a608-aeeb30111be3/volumes" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.711762 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7f857e-08ef-4366-b82c-345bea975113" path="/var/lib/kubelet/pods/2b7f857e-08ef-4366-b82c-345bea975113/volumes" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.712357 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d3762f2-9c1c-40ef-b92b-e7a57a07bf81" path="/var/lib/kubelet/pods/2d3762f2-9c1c-40ef-b92b-e7a57a07bf81/volumes" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.712980 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf5b5257-3852-4cd8-8085-558c8e9bd5d5" path="/var/lib/kubelet/pods/cf5b5257-3852-4cd8-8085-558c8e9bd5d5/volumes" Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.778553 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-4941-account-create-update-7s89w"] Jan 20 09:10:27 crc kubenswrapper[4967]: I0120 09:10:27.929096 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ml9dr"] Jan 20 09:10:27 crc kubenswrapper[4967]: W0120 09:10:27.930442 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f82f406_7f35_4f8b_9b6a_8b5c519a9993.slice/crio-8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074 WatchSource:0}: Error finding container 8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074: Status 404 returned error can't find the container with id 8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074 Jan 20 09:10:28 crc kubenswrapper[4967]: I0120 09:10:28.331088 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" event={"ID":"ed72bd8b-a94b-414c-8162-e500c1d33496","Type":"ContainerStarted","Data":"c72e07f6315df2510f681708b0814775a70aebabc40cc8794fc80b68a358b202"} Jan 20 09:10:28 crc kubenswrapper[4967]: I0120 09:10:28.332166 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" event={"ID":"9f82f406-7f35-4f8b-9b6a-8b5c519a9993","Type":"ContainerStarted","Data":"8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074"} Jan 20 09:10:29 crc kubenswrapper[4967]: I0120 09:10:29.344082 4967 generic.go:334] "Generic (PLEG): container finished" podID="9f82f406-7f35-4f8b-9b6a-8b5c519a9993" containerID="c03e74ad0f638055038a7d4eacc9865025f4db5e78781645cb989600330b1b6f" exitCode=0 Jan 20 09:10:29 crc kubenswrapper[4967]: I0120 09:10:29.344149 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" event={"ID":"9f82f406-7f35-4f8b-9b6a-8b5c519a9993","Type":"ContainerDied","Data":"c03e74ad0f638055038a7d4eacc9865025f4db5e78781645cb989600330b1b6f"} Jan 20 09:10:29 crc kubenswrapper[4967]: I0120 09:10:29.348141 4967 generic.go:334] "Generic (PLEG): container finished" podID="ed72bd8b-a94b-414c-8162-e500c1d33496" containerID="dd2084d9acd3ff737c39b495b30363e0731669a0f84d5dfee73a6fc3aa26e5d5" exitCode=0 Jan 20 09:10:29 crc kubenswrapper[4967]: I0120 09:10:29.348193 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" event={"ID":"ed72bd8b-a94b-414c-8162-e500c1d33496","Type":"ContainerDied","Data":"dd2084d9acd3ff737c39b495b30363e0731669a0f84d5dfee73a6fc3aa26e5d5"} Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.596889 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.661885 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts\") pod \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.661942 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-252ql\" (UniqueName: \"kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql\") pod \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\" (UID: \"9f82f406-7f35-4f8b-9b6a-8b5c519a9993\") " Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.662842 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f82f406-7f35-4f8b-9b6a-8b5c519a9993" (UID: "9f82f406-7f35-4f8b-9b6a-8b5c519a9993"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.667720 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql" (OuterVolumeSpecName: "kube-api-access-252ql") pod "9f82f406-7f35-4f8b-9b6a-8b5c519a9993" (UID: "9f82f406-7f35-4f8b-9b6a-8b5c519a9993"). InnerVolumeSpecName "kube-api-access-252ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.712475 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.762892 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xkdf\" (UniqueName: \"kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf\") pod \"ed72bd8b-a94b-414c-8162-e500c1d33496\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.763019 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts\") pod \"ed72bd8b-a94b-414c-8162-e500c1d33496\" (UID: \"ed72bd8b-a94b-414c-8162-e500c1d33496\") " Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.763412 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.763459 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed72bd8b-a94b-414c-8162-e500c1d33496" (UID: "ed72bd8b-a94b-414c-8162-e500c1d33496"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.763527 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-252ql\" (UniqueName: \"kubernetes.io/projected/9f82f406-7f35-4f8b-9b6a-8b5c519a9993-kube-api-access-252ql\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.765447 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf" (OuterVolumeSpecName: "kube-api-access-8xkdf") pod "ed72bd8b-a94b-414c-8162-e500c1d33496" (UID: "ed72bd8b-a94b-414c-8162-e500c1d33496"). InnerVolumeSpecName "kube-api-access-8xkdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.865235 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed72bd8b-a94b-414c-8162-e500c1d33496-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:30 crc kubenswrapper[4967]: I0120 09:10:30.865288 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xkdf\" (UniqueName: \"kubernetes.io/projected/ed72bd8b-a94b-414c-8162-e500c1d33496-kube-api-access-8xkdf\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.369637 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" event={"ID":"ed72bd8b-a94b-414c-8162-e500c1d33496","Type":"ContainerDied","Data":"c72e07f6315df2510f681708b0814775a70aebabc40cc8794fc80b68a358b202"} Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.369674 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c72e07f6315df2510f681708b0814775a70aebabc40cc8794fc80b68a358b202" Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.369726 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-4941-account-create-update-7s89w" Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.371598 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" event={"ID":"9f82f406-7f35-4f8b-9b6a-8b5c519a9993","Type":"ContainerDied","Data":"8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074"} Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.371731 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ae002c44d4765babf1285e3fd488aafa60eac9c657d66b58349347a810e9074" Jan 20 09:10:31 crc kubenswrapper[4967]: I0120 09:10:31.371783 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-ml9dr" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.758423 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-krm5t"] Jan 20 09:10:32 crc kubenswrapper[4967]: E0120 09:10:32.759025 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed72bd8b-a94b-414c-8162-e500c1d33496" containerName="mariadb-account-create-update" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.759041 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed72bd8b-a94b-414c-8162-e500c1d33496" containerName="mariadb-account-create-update" Jan 20 09:10:32 crc kubenswrapper[4967]: E0120 09:10:32.759068 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f82f406-7f35-4f8b-9b6a-8b5c519a9993" containerName="mariadb-database-create" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.759077 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f82f406-7f35-4f8b-9b6a-8b5c519a9993" containerName="mariadb-database-create" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.759246 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f82f406-7f35-4f8b-9b6a-8b5c519a9993" containerName="mariadb-database-create" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.759262 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed72bd8b-a94b-414c-8162-e500c1d33496" containerName="mariadb-account-create-update" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.759730 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.767251 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.767379 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.767694 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.768371 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-27xq5" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.769228 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-krm5t"] Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.792396 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.792521 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qffc5\" (UniqueName: \"kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.893859 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.894073 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qffc5\" (UniqueName: \"kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.904500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:32 crc kubenswrapper[4967]: I0120 09:10:32.911606 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qffc5\" (UniqueName: \"kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5\") pod \"keystone-db-sync-krm5t\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:33 crc kubenswrapper[4967]: I0120 09:10:33.076418 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:33 crc kubenswrapper[4967]: I0120 09:10:33.271823 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-krm5t"] Jan 20 09:10:33 crc kubenswrapper[4967]: I0120 09:10:33.384904 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" event={"ID":"75d7c33a-b0e4-476a-8ca2-0fb587b63db5","Type":"ContainerStarted","Data":"8fa1672be2813aed49f347901cd2b86c2008b36a9c160ff9767280f026e0b5b7"} Jan 20 09:10:34 crc kubenswrapper[4967]: I0120 09:10:34.406264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" event={"ID":"75d7c33a-b0e4-476a-8ca2-0fb587b63db5","Type":"ContainerStarted","Data":"2012199ae0c330813c9c0e10ae32e0a6ed11e9b92e0f92b36312b89f1cf45850"} Jan 20 09:10:34 crc kubenswrapper[4967]: I0120 09:10:34.423811 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" podStartSLOduration=2.423797712 podStartE2EDuration="2.423797712s" podCreationTimestamp="2026-01-20 09:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:10:34.421380465 +0000 UTC m=+1269.006680672" watchObservedRunningTime="2026-01-20 09:10:34.423797712 +0000 UTC m=+1269.009097919" Jan 20 09:10:35 crc kubenswrapper[4967]: I0120 09:10:35.414574 4967 generic.go:334] "Generic (PLEG): container finished" podID="75d7c33a-b0e4-476a-8ca2-0fb587b63db5" containerID="2012199ae0c330813c9c0e10ae32e0a6ed11e9b92e0f92b36312b89f1cf45850" exitCode=0 Jan 20 09:10:35 crc kubenswrapper[4967]: I0120 09:10:35.414735 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" event={"ID":"75d7c33a-b0e4-476a-8ca2-0fb587b63db5","Type":"ContainerDied","Data":"2012199ae0c330813c9c0e10ae32e0a6ed11e9b92e0f92b36312b89f1cf45850"} Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.674587 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.745142 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data\") pod \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.745250 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qffc5\" (UniqueName: \"kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5\") pod \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\" (UID: \"75d7c33a-b0e4-476a-8ca2-0fb587b63db5\") " Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.750456 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5" (OuterVolumeSpecName: "kube-api-access-qffc5") pod "75d7c33a-b0e4-476a-8ca2-0fb587b63db5" (UID: "75d7c33a-b0e4-476a-8ca2-0fb587b63db5"). InnerVolumeSpecName "kube-api-access-qffc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.776796 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data" (OuterVolumeSpecName: "config-data") pod "75d7c33a-b0e4-476a-8ca2-0fb587b63db5" (UID: "75d7c33a-b0e4-476a-8ca2-0fb587b63db5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.846782 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:36 crc kubenswrapper[4967]: I0120 09:10:36.847011 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qffc5\" (UniqueName: \"kubernetes.io/projected/75d7c33a-b0e4-476a-8ca2-0fb587b63db5-kube-api-access-qffc5\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.429964 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" event={"ID":"75d7c33a-b0e4-476a-8ca2-0fb587b63db5","Type":"ContainerDied","Data":"8fa1672be2813aed49f347901cd2b86c2008b36a9c160ff9767280f026e0b5b7"} Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.430011 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fa1672be2813aed49f347901cd2b86c2008b36a9c160ff9767280f026e0b5b7" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.430026 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-krm5t" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.599978 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-7w867"] Jan 20 09:10:37 crc kubenswrapper[4967]: E0120 09:10:37.600341 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d7c33a-b0e4-476a-8ca2-0fb587b63db5" containerName="keystone-db-sync" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.600360 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d7c33a-b0e4-476a-8ca2-0fb587b63db5" containerName="keystone-db-sync" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.600592 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d7c33a-b0e4-476a-8ca2-0fb587b63db5" containerName="keystone-db-sync" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.601274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.603780 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.604110 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.604143 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.604186 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-27xq5" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.604364 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.621581 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-7w867"] Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.660091 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7stgg\" (UniqueName: \"kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.660136 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.660156 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.660173 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.660200 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.761539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7stgg\" (UniqueName: \"kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.761588 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.761622 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.761639 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.761669 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.766237 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.766816 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.767706 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.781522 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.787333 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7stgg\" (UniqueName: \"kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg\") pod \"keystone-bootstrap-7w867\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:37 crc kubenswrapper[4967]: I0120 09:10:37.915490 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:38 crc kubenswrapper[4967]: I0120 09:10:38.289666 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-7w867"] Jan 20 09:10:38 crc kubenswrapper[4967]: I0120 09:10:38.438924 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" event={"ID":"a473707c-9f85-4475-93b9-adfdc307a5fc","Type":"ContainerStarted","Data":"ce2532cd2d4d6d323ced22f0fbb751a8b077e9a85baf051ed58d836ed424b1f2"} Jan 20 09:10:38 crc kubenswrapper[4967]: I0120 09:10:38.438966 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" event={"ID":"a473707c-9f85-4475-93b9-adfdc307a5fc","Type":"ContainerStarted","Data":"e397d1607eb4ca975a2b8cc606e4861d501ca9c6f51a9209def43a6b9b7ab355"} Jan 20 09:10:38 crc kubenswrapper[4967]: I0120 09:10:38.460673 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" podStartSLOduration=1.460652369 podStartE2EDuration="1.460652369s" podCreationTimestamp="2026-01-20 09:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:10:38.455985381 +0000 UTC m=+1273.041285608" watchObservedRunningTime="2026-01-20 09:10:38.460652369 +0000 UTC m=+1273.045952576" Jan 20 09:10:41 crc kubenswrapper[4967]: I0120 09:10:41.461949 4967 generic.go:334] "Generic (PLEG): container finished" podID="a473707c-9f85-4475-93b9-adfdc307a5fc" containerID="ce2532cd2d4d6d323ced22f0fbb751a8b077e9a85baf051ed58d836ed424b1f2" exitCode=0 Jan 20 09:10:41 crc kubenswrapper[4967]: I0120 09:10:41.462029 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" event={"ID":"a473707c-9f85-4475-93b9-adfdc307a5fc","Type":"ContainerDied","Data":"ce2532cd2d4d6d323ced22f0fbb751a8b077e9a85baf051ed58d836ed424b1f2"} Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.740772 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.931546 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys\") pod \"a473707c-9f85-4475-93b9-adfdc307a5fc\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.931669 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7stgg\" (UniqueName: \"kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg\") pod \"a473707c-9f85-4475-93b9-adfdc307a5fc\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.931759 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts\") pod \"a473707c-9f85-4475-93b9-adfdc307a5fc\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.931782 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys\") pod \"a473707c-9f85-4475-93b9-adfdc307a5fc\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.931846 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data\") pod \"a473707c-9f85-4475-93b9-adfdc307a5fc\" (UID: \"a473707c-9f85-4475-93b9-adfdc307a5fc\") " Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.941305 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts" (OuterVolumeSpecName: "scripts") pod "a473707c-9f85-4475-93b9-adfdc307a5fc" (UID: "a473707c-9f85-4475-93b9-adfdc307a5fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.941317 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg" (OuterVolumeSpecName: "kube-api-access-7stgg") pod "a473707c-9f85-4475-93b9-adfdc307a5fc" (UID: "a473707c-9f85-4475-93b9-adfdc307a5fc"). InnerVolumeSpecName "kube-api-access-7stgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.941379 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a473707c-9f85-4475-93b9-adfdc307a5fc" (UID: "a473707c-9f85-4475-93b9-adfdc307a5fc"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.941596 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a473707c-9f85-4475-93b9-adfdc307a5fc" (UID: "a473707c-9f85-4475-93b9-adfdc307a5fc"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:42 crc kubenswrapper[4967]: I0120 09:10:42.967070 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data" (OuterVolumeSpecName: "config-data") pod "a473707c-9f85-4475-93b9-adfdc307a5fc" (UID: "a473707c-9f85-4475-93b9-adfdc307a5fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.033271 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.033642 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.033653 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.033667 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7stgg\" (UniqueName: \"kubernetes.io/projected/a473707c-9f85-4475-93b9-adfdc307a5fc-kube-api-access-7stgg\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.033680 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a473707c-9f85-4475-93b9-adfdc307a5fc-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.478138 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" event={"ID":"a473707c-9f85-4475-93b9-adfdc307a5fc","Type":"ContainerDied","Data":"e397d1607eb4ca975a2b8cc606e4861d501ca9c6f51a9209def43a6b9b7ab355"} Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.478186 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e397d1607eb4ca975a2b8cc606e4861d501ca9c6f51a9209def43a6b9b7ab355" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.478199 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-7w867" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.552111 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:10:43 crc kubenswrapper[4967]: E0120 09:10:43.552392 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a473707c-9f85-4475-93b9-adfdc307a5fc" containerName="keystone-bootstrap" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.552414 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a473707c-9f85-4475-93b9-adfdc307a5fc" containerName="keystone-bootstrap" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.552559 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a473707c-9f85-4475-93b9-adfdc307a5fc" containerName="keystone-bootstrap" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.553249 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.554933 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.558006 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.558142 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-27xq5" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.558260 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.562238 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.742452 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4k7c\" (UniqueName: \"kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.742511 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.742666 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.742689 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.742722 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.844148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.844216 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.844254 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.844275 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4k7c\" (UniqueName: \"kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.844315 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.848777 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.849491 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.849523 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.849651 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.861409 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4k7c\" (UniqueName: \"kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c\") pod \"keystone-b8c6d96c9-ml4rt\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:43 crc kubenswrapper[4967]: I0120 09:10:43.870302 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:44 crc kubenswrapper[4967]: I0120 09:10:44.280961 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:10:44 crc kubenswrapper[4967]: I0120 09:10:44.660965 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" event={"ID":"62d8ba80-d745-41b8-9044-86ef3fc5264e","Type":"ContainerStarted","Data":"f8488ea83d66bda836d60a57806ef971214b0941f06027ae41cc8938685a4507"} Jan 20 09:10:45 crc kubenswrapper[4967]: I0120 09:10:45.668424 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" event={"ID":"62d8ba80-d745-41b8-9044-86ef3fc5264e","Type":"ContainerStarted","Data":"cd6ad96d6298653d88c39d7c15a7c09d5500610ae262d821a40f8f12b5e5f8e1"} Jan 20 09:10:45 crc kubenswrapper[4967]: I0120 09:10:45.668601 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:10:45 crc kubenswrapper[4967]: I0120 09:10:45.684135 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" podStartSLOduration=2.684117155 podStartE2EDuration="2.684117155s" podCreationTimestamp="2026-01-20 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:10:45.682751317 +0000 UTC m=+1280.268051554" watchObservedRunningTime="2026-01-20 09:10:45.684117155 +0000 UTC m=+1280.269417372" Jan 20 09:11:15 crc kubenswrapper[4967]: I0120 09:11:15.350148 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.011273 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.012238 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.021016 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.023882 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.043294 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.052562 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.104865 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.104935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.104977 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105013 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105046 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105074 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105097 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105132 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6nc8\" (UniqueName: \"kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105155 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8s4s\" (UniqueName: \"kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.105170 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.206642 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.206734 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.206767 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.206861 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.206927 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.207001 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.207065 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.207150 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6nc8\" (UniqueName: \"kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.207181 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8s4s\" (UniqueName: \"kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.207225 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.214399 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.215098 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.215167 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.215678 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.216390 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.221738 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.223154 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.225867 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.227322 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6nc8\" (UniqueName: \"kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8\") pod \"keystone-b8c6d96c9-sm6dk\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.230731 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8s4s\" (UniqueName: \"kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s\") pod \"keystone-b8c6d96c9-rsvvw\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.350925 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.361042 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.660916 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.775915 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:16 crc kubenswrapper[4967]: W0120 09:11:16.781334 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a147513_7ad8_4d63_8117_7538aef75c74.slice/crio-a753cf8b297e2ed6675b776cc44ba9634c962c7a969b1dea2c6b83e1f32d7d4b WatchSource:0}: Error finding container a753cf8b297e2ed6675b776cc44ba9634c962c7a969b1dea2c6b83e1f32d7d4b: Status 404 returned error can't find the container with id a753cf8b297e2ed6675b776cc44ba9634c962c7a969b1dea2c6b83e1f32d7d4b Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.871092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" event={"ID":"5a147513-7ad8-4d63-8117-7538aef75c74","Type":"ContainerStarted","Data":"a753cf8b297e2ed6675b776cc44ba9634c962c7a969b1dea2c6b83e1f32d7d4b"} Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.872753 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" event={"ID":"aed00e46-da2c-4672-8468-dd77b3a304e1","Type":"ContainerStarted","Data":"f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd"} Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.872782 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" event={"ID":"aed00e46-da2c-4672-8468-dd77b3a304e1","Type":"ContainerStarted","Data":"afa5e3a179cbae46d2f3a77fe9fbb81d78c68ed037fd6f804ce9d51d67206b8b"} Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.872924 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:16 crc kubenswrapper[4967]: I0120 09:11:16.889144 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" podStartSLOduration=0.889119781 podStartE2EDuration="889.119781ms" podCreationTimestamp="2026-01-20 09:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:11:16.888520965 +0000 UTC m=+1311.473821192" watchObservedRunningTime="2026-01-20 09:11:16.889119781 +0000 UTC m=+1311.474419988" Jan 20 09:11:17 crc kubenswrapper[4967]: I0120 09:11:17.882199 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" event={"ID":"5a147513-7ad8-4d63-8117-7538aef75c74","Type":"ContainerStarted","Data":"d525b7eac7094e16d8a3baf5c7876e7293b086a1776100568bf3f645ad2ff251"} Jan 20 09:11:17 crc kubenswrapper[4967]: I0120 09:11:17.882646 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:17 crc kubenswrapper[4967]: I0120 09:11:17.903286 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" podStartSLOduration=2.903270113 podStartE2EDuration="2.903270113s" podCreationTimestamp="2026-01-20 09:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:11:17.902260636 +0000 UTC m=+1312.487560853" watchObservedRunningTime="2026-01-20 09:11:17.903270113 +0000 UTC m=+1312.488570320" Jan 20 09:11:48 crc kubenswrapper[4967]: I0120 09:11:48.095529 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:48 crc kubenswrapper[4967]: I0120 09:11:48.169085 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:48 crc kubenswrapper[4967]: I0120 09:11:48.474558 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:11:48 crc kubenswrapper[4967]: I0120 09:11:48.474657 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:11:49 crc kubenswrapper[4967]: I0120 09:11:49.038310 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:49 crc kubenswrapper[4967]: I0120 09:11:49.039204 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" podUID="5a147513-7ad8-4d63-8117-7538aef75c74" containerName="keystone-api" containerID="cri-o://d525b7eac7094e16d8a3baf5c7876e7293b086a1776100568bf3f645ad2ff251" gracePeriod=30 Jan 20 09:11:49 crc kubenswrapper[4967]: I0120 09:11:49.040716 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:49 crc kubenswrapper[4967]: I0120 09:11:49.040927 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" podUID="aed00e46-da2c-4672-8468-dd77b3a304e1" containerName="keystone-api" containerID="cri-o://f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd" gracePeriod=30 Jan 20 09:11:50 crc kubenswrapper[4967]: I0120 09:11:50.260743 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:11:50 crc kubenswrapper[4967]: I0120 09:11:50.266961 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" podUID="62d8ba80-d745-41b8-9044-86ef3fc5264e" containerName="keystone-api" containerID="cri-o://cd6ad96d6298653d88c39d7c15a7c09d5500610ae262d821a40f8f12b5e5f8e1" gracePeriod=30 Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.361653 4967 generic.go:334] "Generic (PLEG): container finished" podID="5a147513-7ad8-4d63-8117-7538aef75c74" containerID="d525b7eac7094e16d8a3baf5c7876e7293b086a1776100568bf3f645ad2ff251" exitCode=0 Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.362073 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" event={"ID":"5a147513-7ad8-4d63-8117-7538aef75c74","Type":"ContainerDied","Data":"d525b7eac7094e16d8a3baf5c7876e7293b086a1776100568bf3f645ad2ff251"} Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.369702 4967 generic.go:334] "Generic (PLEG): container finished" podID="aed00e46-da2c-4672-8468-dd77b3a304e1" containerID="f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd" exitCode=0 Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.369746 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" event={"ID":"aed00e46-da2c-4672-8468-dd77b3a304e1","Type":"ContainerDied","Data":"f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd"} Jan 20 09:11:52 crc kubenswrapper[4967]: E0120 09:11:52.431844 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaed00e46_da2c_4672_8468_dd77b3a304e1.slice/crio-conmon-f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd.scope\": RecentStats: unable to find data in memory cache]" Jan 20 09:11:52 crc kubenswrapper[4967]: E0120 09:11:52.451564 4967 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.558510 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.564756 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744129 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts\") pod \"5a147513-7ad8-4d63-8117-7538aef75c74\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744216 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8s4s\" (UniqueName: \"kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s\") pod \"5a147513-7ad8-4d63-8117-7538aef75c74\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744294 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts\") pod \"aed00e46-da2c-4672-8468-dd77b3a304e1\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744384 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys\") pod \"5a147513-7ad8-4d63-8117-7538aef75c74\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744415 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data\") pod \"aed00e46-da2c-4672-8468-dd77b3a304e1\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.744450 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6nc8\" (UniqueName: \"kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8\") pod \"aed00e46-da2c-4672-8468-dd77b3a304e1\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.745805 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys\") pod \"aed00e46-da2c-4672-8468-dd77b3a304e1\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.745852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys\") pod \"5a147513-7ad8-4d63-8117-7538aef75c74\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.745887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys\") pod \"aed00e46-da2c-4672-8468-dd77b3a304e1\" (UID: \"aed00e46-da2c-4672-8468-dd77b3a304e1\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.745920 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data\") pod \"5a147513-7ad8-4d63-8117-7538aef75c74\" (UID: \"5a147513-7ad8-4d63-8117-7538aef75c74\") " Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.754144 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aed00e46-da2c-4672-8468-dd77b3a304e1" (UID: "aed00e46-da2c-4672-8468-dd77b3a304e1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755720 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s" (OuterVolumeSpecName: "kube-api-access-n8s4s") pod "5a147513-7ad8-4d63-8117-7538aef75c74" (UID: "5a147513-7ad8-4d63-8117-7538aef75c74"). InnerVolumeSpecName "kube-api-access-n8s4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755745 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5a147513-7ad8-4d63-8117-7538aef75c74" (UID: "5a147513-7ad8-4d63-8117-7538aef75c74"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755765 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8" (OuterVolumeSpecName: "kube-api-access-n6nc8") pod "aed00e46-da2c-4672-8468-dd77b3a304e1" (UID: "aed00e46-da2c-4672-8468-dd77b3a304e1"). InnerVolumeSpecName "kube-api-access-n6nc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755725 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts" (OuterVolumeSpecName: "scripts") pod "5a147513-7ad8-4d63-8117-7538aef75c74" (UID: "5a147513-7ad8-4d63-8117-7538aef75c74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755829 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aed00e46-da2c-4672-8468-dd77b3a304e1" (UID: "aed00e46-da2c-4672-8468-dd77b3a304e1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5a147513-7ad8-4d63-8117-7538aef75c74" (UID: "5a147513-7ad8-4d63-8117-7538aef75c74"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.755927 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts" (OuterVolumeSpecName: "scripts") pod "aed00e46-da2c-4672-8468-dd77b3a304e1" (UID: "aed00e46-da2c-4672-8468-dd77b3a304e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.768078 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data" (OuterVolumeSpecName: "config-data") pod "aed00e46-da2c-4672-8468-dd77b3a304e1" (UID: "aed00e46-da2c-4672-8468-dd77b3a304e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.782632 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data" (OuterVolumeSpecName: "config-data") pod "5a147513-7ad8-4d63-8117-7538aef75c74" (UID: "5a147513-7ad8-4d63-8117-7538aef75c74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848272 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848344 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848357 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8s4s\" (UniqueName: \"kubernetes.io/projected/5a147513-7ad8-4d63-8117-7538aef75c74-kube-api-access-n8s4s\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848368 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848375 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848383 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848392 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6nc8\" (UniqueName: \"kubernetes.io/projected/aed00e46-da2c-4672-8468-dd77b3a304e1-kube-api-access-n6nc8\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848399 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848408 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a147513-7ad8-4d63-8117-7538aef75c74-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:52 crc kubenswrapper[4967]: I0120 09:11:52.848416 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aed00e46-da2c-4672-8468-dd77b3a304e1-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.376739 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" event={"ID":"aed00e46-da2c-4672-8468-dd77b3a304e1","Type":"ContainerDied","Data":"afa5e3a179cbae46d2f3a77fe9fbb81d78c68ed037fd6f804ce9d51d67206b8b"} Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.377059 4967 scope.go:117] "RemoveContainer" containerID="f3c5ecc16dbdb056013c909f01c9cd2ab31e7d37da9cf394f085916d8c0c5cdd" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.376752 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.378756 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" event={"ID":"5a147513-7ad8-4d63-8117-7538aef75c74","Type":"ContainerDied","Data":"a753cf8b297e2ed6675b776cc44ba9634c962c7a969b1dea2c6b83e1f32d7d4b"} Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.378894 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.402549 4967 scope.go:117] "RemoveContainer" containerID="d525b7eac7094e16d8a3baf5c7876e7293b086a1776100568bf3f645ad2ff251" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.411172 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.418389 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-sm6dk"] Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.428870 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.433594 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-rsvvw"] Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.703852 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a147513-7ad8-4d63-8117-7538aef75c74" path="/var/lib/kubelet/pods/5a147513-7ad8-4d63-8117-7538aef75c74/volumes" Jan 20 09:11:53 crc kubenswrapper[4967]: I0120 09:11:53.704846 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed00e46-da2c-4672-8468-dd77b3a304e1" path="/var/lib/kubelet/pods/aed00e46-da2c-4672-8468-dd77b3a304e1/volumes" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.385889 4967 generic.go:334] "Generic (PLEG): container finished" podID="62d8ba80-d745-41b8-9044-86ef3fc5264e" containerID="cd6ad96d6298653d88c39d7c15a7c09d5500610ae262d821a40f8f12b5e5f8e1" exitCode=0 Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.385963 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" event={"ID":"62d8ba80-d745-41b8-9044-86ef3fc5264e","Type":"ContainerDied","Data":"cd6ad96d6298653d88c39d7c15a7c09d5500610ae262d821a40f8f12b5e5f8e1"} Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.546398 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.672762 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys\") pod \"62d8ba80-d745-41b8-9044-86ef3fc5264e\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.672846 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4k7c\" (UniqueName: \"kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c\") pod \"62d8ba80-d745-41b8-9044-86ef3fc5264e\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.672899 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys\") pod \"62d8ba80-d745-41b8-9044-86ef3fc5264e\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.672919 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data\") pod \"62d8ba80-d745-41b8-9044-86ef3fc5264e\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.672956 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts\") pod \"62d8ba80-d745-41b8-9044-86ef3fc5264e\" (UID: \"62d8ba80-d745-41b8-9044-86ef3fc5264e\") " Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.678182 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "62d8ba80-d745-41b8-9044-86ef3fc5264e" (UID: "62d8ba80-d745-41b8-9044-86ef3fc5264e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.678347 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts" (OuterVolumeSpecName: "scripts") pod "62d8ba80-d745-41b8-9044-86ef3fc5264e" (UID: "62d8ba80-d745-41b8-9044-86ef3fc5264e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.678502 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c" (OuterVolumeSpecName: "kube-api-access-c4k7c") pod "62d8ba80-d745-41b8-9044-86ef3fc5264e" (UID: "62d8ba80-d745-41b8-9044-86ef3fc5264e"). InnerVolumeSpecName "kube-api-access-c4k7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.680791 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "62d8ba80-d745-41b8-9044-86ef3fc5264e" (UID: "62d8ba80-d745-41b8-9044-86ef3fc5264e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.692013 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data" (OuterVolumeSpecName: "config-data") pod "62d8ba80-d745-41b8-9044-86ef3fc5264e" (UID: "62d8ba80-d745-41b8-9044-86ef3fc5264e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.776249 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4k7c\" (UniqueName: \"kubernetes.io/projected/62d8ba80-d745-41b8-9044-86ef3fc5264e-kube-api-access-c4k7c\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.776313 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.776340 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.776363 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:54 crc kubenswrapper[4967]: I0120 09:11:54.776386 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62d8ba80-d745-41b8-9044-86ef3fc5264e-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.396154 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" event={"ID":"62d8ba80-d745-41b8-9044-86ef3fc5264e","Type":"ContainerDied","Data":"f8488ea83d66bda836d60a57806ef971214b0941f06027ae41cc8938685a4507"} Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.396193 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt" Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.396215 4967 scope.go:117] "RemoveContainer" containerID="cd6ad96d6298653d88c39d7c15a7c09d5500610ae262d821a40f8f12b5e5f8e1" Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.427947 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.432302 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-b8c6d96c9-ml4rt"] Jan 20 09:11:55 crc kubenswrapper[4967]: I0120 09:11:55.700853 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d8ba80-d745-41b8-9044-86ef3fc5264e" path="/var/lib/kubelet/pods/62d8ba80-d745-41b8-9044-86ef3fc5264e/volumes" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.428539 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-krm5t"] Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.432713 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-krm5t"] Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.448510 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-7w867"] Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.454288 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-7w867"] Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.500568 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone4941-account-delete-llzf2"] Jan 20 09:11:56 crc kubenswrapper[4967]: E0120 09:11:56.500842 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed00e46-da2c-4672-8468-dd77b3a304e1" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.500856 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed00e46-da2c-4672-8468-dd77b3a304e1" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: E0120 09:11:56.500865 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d8ba80-d745-41b8-9044-86ef3fc5264e" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.500870 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d8ba80-d745-41b8-9044-86ef3fc5264e" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: E0120 09:11:56.500883 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a147513-7ad8-4d63-8117-7538aef75c74" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.500889 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a147513-7ad8-4d63-8117-7538aef75c74" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.500997 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d8ba80-d745-41b8-9044-86ef3fc5264e" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.501008 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed00e46-da2c-4672-8468-dd77b3a304e1" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.501017 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a147513-7ad8-4d63-8117-7538aef75c74" containerName="keystone-api" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.501444 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.510132 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone4941-account-delete-llzf2"] Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.703952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nkqx\" (UniqueName: \"kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.704097 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.805074 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.805183 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nkqx\" (UniqueName: \"kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.805894 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:56 crc kubenswrapper[4967]: I0120 09:11:56.826473 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nkqx\" (UniqueName: \"kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx\") pod \"keystone4941-account-delete-llzf2\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:57 crc kubenswrapper[4967]: I0120 09:11:57.117220 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:57 crc kubenswrapper[4967]: I0120 09:11:57.533199 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone4941-account-delete-llzf2"] Jan 20 09:11:57 crc kubenswrapper[4967]: I0120 09:11:57.702655 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d7c33a-b0e4-476a-8ca2-0fb587b63db5" path="/var/lib/kubelet/pods/75d7c33a-b0e4-476a-8ca2-0fb587b63db5/volumes" Jan 20 09:11:57 crc kubenswrapper[4967]: I0120 09:11:57.703488 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a473707c-9f85-4475-93b9-adfdc307a5fc" path="/var/lib/kubelet/pods/a473707c-9f85-4475-93b9-adfdc307a5fc/volumes" Jan 20 09:11:58 crc kubenswrapper[4967]: I0120 09:11:58.419329 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3f3b953-ee43-45f7-a405-d1bf15a98556" containerID="3f4434d588871cdbeb73529bf7aea288ba1bad8370ad11029c6fe8bb7580ad0d" exitCode=0 Jan 20 09:11:58 crc kubenswrapper[4967]: I0120 09:11:58.419526 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" event={"ID":"f3f3b953-ee43-45f7-a405-d1bf15a98556","Type":"ContainerDied","Data":"3f4434d588871cdbeb73529bf7aea288ba1bad8370ad11029c6fe8bb7580ad0d"} Jan 20 09:11:58 crc kubenswrapper[4967]: I0120 09:11:58.419601 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" event={"ID":"f3f3b953-ee43-45f7-a405-d1bf15a98556","Type":"ContainerStarted","Data":"786a65a066810fb9240d2a6c9c8552deade6e92f9a5ffe2ea66267d71c0c1a2c"} Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.655280 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.842663 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts\") pod \"f3f3b953-ee43-45f7-a405-d1bf15a98556\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.842748 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nkqx\" (UniqueName: \"kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx\") pod \"f3f3b953-ee43-45f7-a405-d1bf15a98556\" (UID: \"f3f3b953-ee43-45f7-a405-d1bf15a98556\") " Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.843538 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f3f3b953-ee43-45f7-a405-d1bf15a98556" (UID: "f3f3b953-ee43-45f7-a405-d1bf15a98556"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.847665 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx" (OuterVolumeSpecName: "kube-api-access-6nkqx") pod "f3f3b953-ee43-45f7-a405-d1bf15a98556" (UID: "f3f3b953-ee43-45f7-a405-d1bf15a98556"). InnerVolumeSpecName "kube-api-access-6nkqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.944466 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3f3b953-ee43-45f7-a405-d1bf15a98556-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:11:59 crc kubenswrapper[4967]: I0120 09:11:59.944529 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nkqx\" (UniqueName: \"kubernetes.io/projected/f3f3b953-ee43-45f7-a405-d1bf15a98556-kube-api-access-6nkqx\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:00 crc kubenswrapper[4967]: I0120 09:12:00.435310 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" event={"ID":"f3f3b953-ee43-45f7-a405-d1bf15a98556","Type":"ContainerDied","Data":"786a65a066810fb9240d2a6c9c8552deade6e92f9a5ffe2ea66267d71c0c1a2c"} Jan 20 09:12:00 crc kubenswrapper[4967]: I0120 09:12:00.435356 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786a65a066810fb9240d2a6c9c8552deade6e92f9a5ffe2ea66267d71c0c1a2c" Jan 20 09:12:00 crc kubenswrapper[4967]: I0120 09:12:00.435383 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone4941-account-delete-llzf2" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.522921 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ml9dr"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.527474 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-ml9dr"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.543373 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone4941-account-delete-llzf2"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.561739 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone4941-account-delete-llzf2"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.568052 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-4941-account-create-update-7s89w"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.574069 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-4941-account-create-update-7s89w"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.656036 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-8g5fd"] Jan 20 09:12:01 crc kubenswrapper[4967]: E0120 09:12:01.656341 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3f3b953-ee43-45f7-a405-d1bf15a98556" containerName="mariadb-account-delete" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.656355 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3f3b953-ee43-45f7-a405-d1bf15a98556" containerName="mariadb-account-delete" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.656495 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3f3b953-ee43-45f7-a405-d1bf15a98556" containerName="mariadb-account-delete" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.657092 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.661324 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-8g5fd"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.705335 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f82f406-7f35-4f8b-9b6a-8b5c519a9993" path="/var/lib/kubelet/pods/9f82f406-7f35-4f8b-9b6a-8b5c519a9993/volumes" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.706406 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed72bd8b-a94b-414c-8162-e500c1d33496" path="/var/lib/kubelet/pods/ed72bd8b-a94b-414c-8162-e500c1d33496/volumes" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.707297 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3f3b953-ee43-45f7-a405-d1bf15a98556" path="/var/lib/kubelet/pods/f3f3b953-ee43-45f7-a405-d1bf15a98556/volumes" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.753735 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.754586 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.756321 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.762975 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg"] Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.775399 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.775682 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlmxk\" (UniqueName: \"kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.877268 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlmxk\" (UniqueName: \"kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.877399 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.877425 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.877478 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n62s\" (UniqueName: \"kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.878801 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.894170 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlmxk\" (UniqueName: \"kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk\") pod \"keystone-db-create-8g5fd\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.979318 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.979427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n62s\" (UniqueName: \"kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.980535 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.981890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:01 crc kubenswrapper[4967]: I0120 09:12:01.996643 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n62s\" (UniqueName: \"kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s\") pod \"keystone-f12d-account-create-update-k2wkg\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.069468 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.292407 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg"] Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.414808 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-8g5fd"] Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.449051 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" event={"ID":"dc3ef008-51cc-4810-89b7-507f156ed524","Type":"ContainerStarted","Data":"2af072d4195b894d117a42c260d7e2909a983afe4638c2ccfa8358b2731e581d"} Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.449155 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" event={"ID":"dc3ef008-51cc-4810-89b7-507f156ed524","Type":"ContainerStarted","Data":"ae59924b3123fd4a88a990dbde4af5c527a1f30a4330ed1b3b7afbd25ceac15b"} Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.451255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" event={"ID":"3287d1c7-fc4d-4449-ad93-369d4631dc72","Type":"ContainerStarted","Data":"de12f5cb141138106947191743ddc1d0bc6818441dcc54165154c637bf467e10"} Jan 20 09:12:02 crc kubenswrapper[4967]: I0120 09:12:02.466522 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" podStartSLOduration=1.466503042 podStartE2EDuration="1.466503042s" podCreationTimestamp="2026-01-20 09:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:12:02.462351893 +0000 UTC m=+1357.047652140" watchObservedRunningTime="2026-01-20 09:12:02.466503042 +0000 UTC m=+1357.051803269" Jan 20 09:12:03 crc kubenswrapper[4967]: I0120 09:12:03.459470 4967 generic.go:334] "Generic (PLEG): container finished" podID="3287d1c7-fc4d-4449-ad93-369d4631dc72" containerID="eff3a2e2f9d176b6a11fb1e700aa1d2d3dfe26ef48446d0193fef0f1c8ced2c0" exitCode=0 Jan 20 09:12:03 crc kubenswrapper[4967]: I0120 09:12:03.459495 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" event={"ID":"3287d1c7-fc4d-4449-ad93-369d4631dc72","Type":"ContainerDied","Data":"eff3a2e2f9d176b6a11fb1e700aa1d2d3dfe26ef48446d0193fef0f1c8ced2c0"} Jan 20 09:12:03 crc kubenswrapper[4967]: I0120 09:12:03.461322 4967 generic.go:334] "Generic (PLEG): container finished" podID="dc3ef008-51cc-4810-89b7-507f156ed524" containerID="2af072d4195b894d117a42c260d7e2909a983afe4638c2ccfa8358b2731e581d" exitCode=0 Jan 20 09:12:03 crc kubenswrapper[4967]: I0120 09:12:03.461363 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" event={"ID":"dc3ef008-51cc-4810-89b7-507f156ed524","Type":"ContainerDied","Data":"2af072d4195b894d117a42c260d7e2909a983afe4638c2ccfa8358b2731e581d"} Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.780651 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.786124 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.924293 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlmxk\" (UniqueName: \"kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk\") pod \"3287d1c7-fc4d-4449-ad93-369d4631dc72\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.924339 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n62s\" (UniqueName: \"kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s\") pod \"dc3ef008-51cc-4810-89b7-507f156ed524\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.924369 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts\") pod \"dc3ef008-51cc-4810-89b7-507f156ed524\" (UID: \"dc3ef008-51cc-4810-89b7-507f156ed524\") " Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.924505 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts\") pod \"3287d1c7-fc4d-4449-ad93-369d4631dc72\" (UID: \"3287d1c7-fc4d-4449-ad93-369d4631dc72\") " Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.925280 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc3ef008-51cc-4810-89b7-507f156ed524" (UID: "dc3ef008-51cc-4810-89b7-507f156ed524"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.925291 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3287d1c7-fc4d-4449-ad93-369d4631dc72" (UID: "3287d1c7-fc4d-4449-ad93-369d4631dc72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.930335 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s" (OuterVolumeSpecName: "kube-api-access-4n62s") pod "dc3ef008-51cc-4810-89b7-507f156ed524" (UID: "dc3ef008-51cc-4810-89b7-507f156ed524"). InnerVolumeSpecName "kube-api-access-4n62s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:04 crc kubenswrapper[4967]: I0120 09:12:04.930842 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk" (OuterVolumeSpecName: "kube-api-access-rlmxk") pod "3287d1c7-fc4d-4449-ad93-369d4631dc72" (UID: "3287d1c7-fc4d-4449-ad93-369d4631dc72"). InnerVolumeSpecName "kube-api-access-rlmxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.025811 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlmxk\" (UniqueName: \"kubernetes.io/projected/3287d1c7-fc4d-4449-ad93-369d4631dc72-kube-api-access-rlmxk\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.025862 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n62s\" (UniqueName: \"kubernetes.io/projected/dc3ef008-51cc-4810-89b7-507f156ed524-kube-api-access-4n62s\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.025880 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc3ef008-51cc-4810-89b7-507f156ed524-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.025897 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3287d1c7-fc4d-4449-ad93-369d4631dc72-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.487776 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" event={"ID":"dc3ef008-51cc-4810-89b7-507f156ed524","Type":"ContainerDied","Data":"ae59924b3123fd4a88a990dbde4af5c527a1f30a4330ed1b3b7afbd25ceac15b"} Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.487811 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.487818 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae59924b3123fd4a88a990dbde4af5c527a1f30a4330ed1b3b7afbd25ceac15b" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.489544 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" event={"ID":"3287d1c7-fc4d-4449-ad93-369d4631dc72","Type":"ContainerDied","Data":"de12f5cb141138106947191743ddc1d0bc6818441dcc54165154c637bf467e10"} Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.489582 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de12f5cb141138106947191743ddc1d0bc6818441dcc54165154c637bf467e10" Jan 20 09:12:05 crc kubenswrapper[4967]: I0120 09:12:05.489717 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-8g5fd" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.319196 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-5m5dg"] Jan 20 09:12:07 crc kubenswrapper[4967]: E0120 09:12:07.319813 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3287d1c7-fc4d-4449-ad93-369d4631dc72" containerName="mariadb-database-create" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.319833 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3287d1c7-fc4d-4449-ad93-369d4631dc72" containerName="mariadb-database-create" Jan 20 09:12:07 crc kubenswrapper[4967]: E0120 09:12:07.319859 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3ef008-51cc-4810-89b7-507f156ed524" containerName="mariadb-account-create-update" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.319871 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3ef008-51cc-4810-89b7-507f156ed524" containerName="mariadb-account-create-update" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.320031 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3287d1c7-fc4d-4449-ad93-369d4631dc72" containerName="mariadb-database-create" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.320051 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3ef008-51cc-4810-89b7-507f156ed524" containerName="mariadb-account-create-update" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.320600 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.325178 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.325234 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vnndf" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.325394 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.325452 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.325502 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.334222 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-5m5dg"] Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.456245 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.456284 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d8rk\" (UniqueName: \"kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.456313 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.557518 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.557571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d8rk\" (UniqueName: \"kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.557635 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.566422 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.574319 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d8rk\" (UniqueName: \"kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.578558 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data\") pod \"keystone-db-sync-5m5dg\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:07 crc kubenswrapper[4967]: I0120 09:12:07.645690 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:08 crc kubenswrapper[4967]: I0120 09:12:08.051946 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-5m5dg"] Jan 20 09:12:08 crc kubenswrapper[4967]: I0120 09:12:08.521073 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" event={"ID":"95d1e0c2-5441-4f92-9058-94bf7eb373db","Type":"ContainerStarted","Data":"c878c6bd96ecf98696e86ce007b6150e1ae1801a931f59eba063a3d3efbc5d9b"} Jan 20 09:12:08 crc kubenswrapper[4967]: I0120 09:12:08.521136 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" event={"ID":"95d1e0c2-5441-4f92-9058-94bf7eb373db","Type":"ContainerStarted","Data":"aed80deedbc67fe583fdd829bcbd26a53146f2273e9cb2cb1ecc03591e10781b"} Jan 20 09:12:08 crc kubenswrapper[4967]: I0120 09:12:08.537212 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" podStartSLOduration=1.537193411 podStartE2EDuration="1.537193411s" podCreationTimestamp="2026-01-20 09:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:12:08.534213673 +0000 UTC m=+1363.119513880" watchObservedRunningTime="2026-01-20 09:12:08.537193411 +0000 UTC m=+1363.122493618" Jan 20 09:12:10 crc kubenswrapper[4967]: I0120 09:12:10.535696 4967 generic.go:334] "Generic (PLEG): container finished" podID="95d1e0c2-5441-4f92-9058-94bf7eb373db" containerID="c878c6bd96ecf98696e86ce007b6150e1ae1801a931f59eba063a3d3efbc5d9b" exitCode=0 Jan 20 09:12:10 crc kubenswrapper[4967]: I0120 09:12:10.535774 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" event={"ID":"95d1e0c2-5441-4f92-9058-94bf7eb373db","Type":"ContainerDied","Data":"c878c6bd96ecf98696e86ce007b6150e1ae1801a931f59eba063a3d3efbc5d9b"} Jan 20 09:12:11 crc kubenswrapper[4967]: I0120 09:12:11.828046 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.022094 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d8rk\" (UniqueName: \"kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk\") pod \"95d1e0c2-5441-4f92-9058-94bf7eb373db\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.022152 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle\") pod \"95d1e0c2-5441-4f92-9058-94bf7eb373db\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.022194 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data\") pod \"95d1e0c2-5441-4f92-9058-94bf7eb373db\" (UID: \"95d1e0c2-5441-4f92-9058-94bf7eb373db\") " Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.038488 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk" (OuterVolumeSpecName: "kube-api-access-4d8rk") pod "95d1e0c2-5441-4f92-9058-94bf7eb373db" (UID: "95d1e0c2-5441-4f92-9058-94bf7eb373db"). InnerVolumeSpecName "kube-api-access-4d8rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.040832 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95d1e0c2-5441-4f92-9058-94bf7eb373db" (UID: "95d1e0c2-5441-4f92-9058-94bf7eb373db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.074931 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data" (OuterVolumeSpecName: "config-data") pod "95d1e0c2-5441-4f92-9058-94bf7eb373db" (UID: "95d1e0c2-5441-4f92-9058-94bf7eb373db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.123866 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d8rk\" (UniqueName: \"kubernetes.io/projected/95d1e0c2-5441-4f92-9058-94bf7eb373db-kube-api-access-4d8rk\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.124032 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.124090 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d1e0c2-5441-4f92-9058-94bf7eb373db-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.553984 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" event={"ID":"95d1e0c2-5441-4f92-9058-94bf7eb373db","Type":"ContainerDied","Data":"aed80deedbc67fe583fdd829bcbd26a53146f2273e9cb2cb1ecc03591e10781b"} Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.554040 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aed80deedbc67fe583fdd829bcbd26a53146f2273e9cb2cb1ecc03591e10781b" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.554094 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-5m5dg" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.736794 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-r4kmx"] Jan 20 09:12:12 crc kubenswrapper[4967]: E0120 09:12:12.737319 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d1e0c2-5441-4f92-9058-94bf7eb373db" containerName="keystone-db-sync" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.737398 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d1e0c2-5441-4f92-9058-94bf7eb373db" containerName="keystone-db-sync" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.737584 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d1e0c2-5441-4f92-9058-94bf7eb373db" containerName="keystone-db-sync" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.738168 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.742595 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.742773 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.742837 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.743018 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.743141 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.743251 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vnndf" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.747246 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-r4kmx"] Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938281 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938577 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938604 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938802 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xplnc\" (UniqueName: \"kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938876 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:12 crc kubenswrapper[4967]: I0120 09:12:12.938923 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.040195 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.041020 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.041102 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.041137 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.041217 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.041273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xplnc\" (UniqueName: \"kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.044442 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.045085 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.046814 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.047037 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.049121 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.070341 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xplnc\" (UniqueName: \"kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc\") pod \"keystone-bootstrap-r4kmx\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.086248 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.316140 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-r4kmx"] Jan 20 09:12:13 crc kubenswrapper[4967]: I0120 09:12:13.563019 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" event={"ID":"f7e7113b-179d-44bc-b663-6de8bcca41c3","Type":"ContainerStarted","Data":"a02a463763d6d636881d527c179c35015af0964588d59f5fc8fb63e289a81032"} Jan 20 09:12:14 crc kubenswrapper[4967]: I0120 09:12:14.570728 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" event={"ID":"f7e7113b-179d-44bc-b663-6de8bcca41c3","Type":"ContainerStarted","Data":"0bd1d7fe79bc8def86b98d6b9d9262b7287d0c38474b0e39e3784e72af8f5504"} Jan 20 09:12:14 crc kubenswrapper[4967]: I0120 09:12:14.593634 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" podStartSLOduration=2.593590221 podStartE2EDuration="2.593590221s" podCreationTimestamp="2026-01-20 09:12:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:12:14.588076976 +0000 UTC m=+1369.173377203" watchObservedRunningTime="2026-01-20 09:12:14.593590221 +0000 UTC m=+1369.178890428" Jan 20 09:12:16 crc kubenswrapper[4967]: I0120 09:12:16.588074 4967 generic.go:334] "Generic (PLEG): container finished" podID="f7e7113b-179d-44bc-b663-6de8bcca41c3" containerID="0bd1d7fe79bc8def86b98d6b9d9262b7287d0c38474b0e39e3784e72af8f5504" exitCode=0 Jan 20 09:12:16 crc kubenswrapper[4967]: I0120 09:12:16.588200 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" event={"ID":"f7e7113b-179d-44bc-b663-6de8bcca41c3","Type":"ContainerDied","Data":"0bd1d7fe79bc8def86b98d6b9d9262b7287d0c38474b0e39e3784e72af8f5504"} Jan 20 09:12:17 crc kubenswrapper[4967]: I0120 09:12:17.909095 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016628 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016665 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016773 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xplnc\" (UniqueName: \"kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016798 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016824 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.016848 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts\") pod \"f7e7113b-179d-44bc-b663-6de8bcca41c3\" (UID: \"f7e7113b-179d-44bc-b663-6de8bcca41c3\") " Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.022637 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc" (OuterVolumeSpecName: "kube-api-access-xplnc") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "kube-api-access-xplnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.023126 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts" (OuterVolumeSpecName: "scripts") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.026977 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.029758 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.035373 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.042957 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data" (OuterVolumeSpecName: "config-data") pod "f7e7113b-179d-44bc-b663-6de8bcca41c3" (UID: "f7e7113b-179d-44bc-b663-6de8bcca41c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.125933 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xplnc\" (UniqueName: \"kubernetes.io/projected/f7e7113b-179d-44bc-b663-6de8bcca41c3-kube-api-access-xplnc\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.125980 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.126000 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.126011 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.126022 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.126033 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7e7113b-179d-44bc-b663-6de8bcca41c3-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.474670 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.474751 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.605375 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" event={"ID":"f7e7113b-179d-44bc-b663-6de8bcca41c3","Type":"ContainerDied","Data":"a02a463763d6d636881d527c179c35015af0964588d59f5fc8fb63e289a81032"} Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.605415 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-r4kmx" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.605417 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a02a463763d6d636881d527c179c35015af0964588d59f5fc8fb63e289a81032" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.804792 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:18 crc kubenswrapper[4967]: E0120 09:12:18.805110 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e7113b-179d-44bc-b663-6de8bcca41c3" containerName="keystone-bootstrap" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.805138 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e7113b-179d-44bc-b663-6de8bcca41c3" containerName="keystone-bootstrap" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.805351 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e7113b-179d-44bc-b663-6de8bcca41c3" containerName="keystone-bootstrap" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.805951 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.808346 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.808603 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.808635 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"cert-keystone-public-svc" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.808676 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"combined-ca-bundle" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.809034 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-vnndf" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.809135 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"cert-keystone-internal-svc" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.810366 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.819008 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.836894 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.836968 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.836989 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wczp2\" (UniqueName: \"kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.837008 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.837025 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.837066 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.837082 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.837098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.938656 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.938993 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939041 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939135 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939211 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939240 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wczp2\" (UniqueName: \"kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939278 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.939310 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.944314 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.944510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.954771 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.955281 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.955966 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.956478 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.957511 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:18 crc kubenswrapper[4967]: I0120 09:12:18.959820 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wczp2\" (UniqueName: \"kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2\") pod \"keystone-76fd8f9b94-4gwss\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.122013 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.321852 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.613157 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" event={"ID":"558871cd-7ca8-4399-a7d7-346605104bb1","Type":"ContainerStarted","Data":"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338"} Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.613633 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.613673 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" event={"ID":"558871cd-7ca8-4399-a7d7-346605104bb1","Type":"ContainerStarted","Data":"1224468d6c19bceb02b17e16c27fb70d84699fa0c03f4515dfddd8f2706e7af7"} Jan 20 09:12:19 crc kubenswrapper[4967]: I0120 09:12:19.631049 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" podStartSLOduration=1.6310319340000001 podStartE2EDuration="1.631031934s" podCreationTimestamp="2026-01-20 09:12:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:12:19.630451928 +0000 UTC m=+1374.215752175" watchObservedRunningTime="2026-01-20 09:12:19.631031934 +0000 UTC m=+1374.216332141" Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.474316 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.474905 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.474958 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.475656 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.475847 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e" gracePeriod=600 Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.846833 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e" exitCode=0 Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.846957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e"} Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.847196 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2"} Jan 20 09:12:48 crc kubenswrapper[4967]: I0120 09:12:48.847214 4967 scope.go:117] "RemoveContainer" containerID="d0d8e79a419e5da544f6c60b84a73fbbc2e45cb4601b43a2c0aac62e41ffcec3" Jan 20 09:12:50 crc kubenswrapper[4967]: I0120 09:12:50.630997 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.354389 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-5m5dg"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.370678 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-r4kmx"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.375448 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-5m5dg"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.382461 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-r4kmx"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.387915 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.388098 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" podUID="558871cd-7ca8-4399-a7d7-346605104bb1" containerName="keystone-api" containerID="cri-o://5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338" gracePeriod=30 Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.432477 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystonef12d-account-delete-gqs4c"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.433441 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.442018 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonef12d-account-delete-gqs4c"] Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.618652 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.618693 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh5gw\" (UniqueName: \"kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.706092 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d1e0c2-5441-4f92-9058-94bf7eb373db" path="/var/lib/kubelet/pods/95d1e0c2-5441-4f92-9058-94bf7eb373db/volumes" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.708136 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e7113b-179d-44bc-b663-6de8bcca41c3" path="/var/lib/kubelet/pods/f7e7113b-179d-44bc-b663-6de8bcca41c3/volumes" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.719811 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.719851 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh5gw\" (UniqueName: \"kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.720768 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:51 crc kubenswrapper[4967]: I0120 09:12:51.765532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh5gw\" (UniqueName: \"kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw\") pod \"keystonef12d-account-delete-gqs4c\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:52 crc kubenswrapper[4967]: I0120 09:12:52.053679 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:52 crc kubenswrapper[4967]: I0120 09:12:52.269146 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonef12d-account-delete-gqs4c"] Jan 20 09:12:52 crc kubenswrapper[4967]: I0120 09:12:52.875790 4967 generic.go:334] "Generic (PLEG): container finished" podID="e009e9d1-fa56-4174-bdd0-29b5e2974b95" containerID="2e1fefd0f41067fcb00e4843ac0f79abf3545a40d39bb1f9a31a6d11939fa584" exitCode=0 Jan 20 09:12:52 crc kubenswrapper[4967]: I0120 09:12:52.875827 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" event={"ID":"e009e9d1-fa56-4174-bdd0-29b5e2974b95","Type":"ContainerDied","Data":"2e1fefd0f41067fcb00e4843ac0f79abf3545a40d39bb1f9a31a6d11939fa584"} Jan 20 09:12:52 crc kubenswrapper[4967]: I0120 09:12:52.876390 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" event={"ID":"e009e9d1-fa56-4174-bdd0-29b5e2974b95","Type":"ContainerStarted","Data":"e473a23583dc7311a3705c2c4b3f1755ed134047c978677478713d788b1e3fcf"} Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.170683 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.255066 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh5gw\" (UniqueName: \"kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw\") pod \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.255238 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts\") pod \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\" (UID: \"e009e9d1-fa56-4174-bdd0-29b5e2974b95\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.256085 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e009e9d1-fa56-4174-bdd0-29b5e2974b95" (UID: "e009e9d1-fa56-4174-bdd0-29b5e2974b95"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.262854 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw" (OuterVolumeSpecName: "kube-api-access-fh5gw") pod "e009e9d1-fa56-4174-bdd0-29b5e2974b95" (UID: "e009e9d1-fa56-4174-bdd0-29b5e2974b95"). InnerVolumeSpecName "kube-api-access-fh5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.357168 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh5gw\" (UniqueName: \"kubernetes.io/projected/e009e9d1-fa56-4174-bdd0-29b5e2974b95-kube-api-access-fh5gw\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.357198 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e009e9d1-fa56-4174-bdd0-29b5e2974b95-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.844380 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.889399 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" event={"ID":"e009e9d1-fa56-4174-bdd0-29b5e2974b95","Type":"ContainerDied","Data":"e473a23583dc7311a3705c2c4b3f1755ed134047c978677478713d788b1e3fcf"} Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.889420 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef12d-account-delete-gqs4c" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.889432 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e473a23583dc7311a3705c2c4b3f1755ed134047c978677478713d788b1e3fcf" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.890882 4967 generic.go:334] "Generic (PLEG): container finished" podID="558871cd-7ca8-4399-a7d7-346605104bb1" containerID="5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338" exitCode=0 Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.890909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" event={"ID":"558871cd-7ca8-4399-a7d7-346605104bb1","Type":"ContainerDied","Data":"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338"} Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.890924 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" event={"ID":"558871cd-7ca8-4399-a7d7-346605104bb1","Type":"ContainerDied","Data":"1224468d6c19bceb02b17e16c27fb70d84699fa0c03f4515dfddd8f2706e7af7"} Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.890939 4967 scope.go:117] "RemoveContainer" containerID="5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.891031 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-76fd8f9b94-4gwss" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.918238 4967 scope.go:117] "RemoveContainer" containerID="5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338" Jan 20 09:12:54 crc kubenswrapper[4967]: E0120 09:12:54.918695 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338\": container with ID starting with 5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338 not found: ID does not exist" containerID="5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.918736 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338"} err="failed to get container status \"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338\": rpc error: code = NotFound desc = could not find container \"5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338\": container with ID starting with 5d56108c62c5b1493752ebba21c9f1033a1a291ef066e7b2b59bcb5435e9f338 not found: ID does not exist" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970236 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970295 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970321 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970381 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970434 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970457 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970525 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.970584 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wczp2\" (UniqueName: \"kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.974086 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.974585 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2" (OuterVolumeSpecName: "kube-api-access-wczp2") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "kube-api-access-wczp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.974734 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.975335 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts" (OuterVolumeSpecName: "scripts") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.987939 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:54 crc kubenswrapper[4967]: I0120 09:12:54.998729 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data" (OuterVolumeSpecName: "config-data") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:55 crc kubenswrapper[4967]: E0120 09:12:55.000925 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs podName:558871cd-7ca8-4399-a7d7-346605104bb1 nodeName:}" failed. No retries permitted until 2026-01-20 09:12:55.500902621 +0000 UTC m=+1410.086202828 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "public-tls-certs" (UniqueName: "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1") : error deleting /var/lib/kubelet/pods/558871cd-7ca8-4399-a7d7-346605104bb1/volume-subpaths: remove /var/lib/kubelet/pods/558871cd-7ca8-4399-a7d7-346605104bb1/volume-subpaths: no such file or directory Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.003581 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072669 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072712 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wczp2\" (UniqueName: \"kubernetes.io/projected/558871cd-7ca8-4399-a7d7-346605104bb1-kube-api-access-wczp2\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072723 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072731 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072740 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072750 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.072758 4967 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.579947 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") pod \"558871cd-7ca8-4399-a7d7-346605104bb1\" (UID: \"558871cd-7ca8-4399-a7d7-346605104bb1\") " Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.601835 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "558871cd-7ca8-4399-a7d7-346605104bb1" (UID: "558871cd-7ca8-4399-a7d7-346605104bb1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.681952 4967 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/558871cd-7ca8-4399-a7d7-346605104bb1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.806024 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:55 crc kubenswrapper[4967]: I0120 09:12:55.812455 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-76fd8f9b94-4gwss"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.451115 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-8g5fd"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.456031 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-8g5fd"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.462697 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystonef12d-account-delete-gqs4c"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.467701 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystonef12d-account-delete-gqs4c"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.471859 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.475795 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-f12d-account-create-update-k2wkg"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.735921 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-slwx4"] Jan 20 09:12:56 crc kubenswrapper[4967]: E0120 09:12:56.736275 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e009e9d1-fa56-4174-bdd0-29b5e2974b95" containerName="mariadb-account-delete" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.736292 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e009e9d1-fa56-4174-bdd0-29b5e2974b95" containerName="mariadb-account-delete" Jan 20 09:12:56 crc kubenswrapper[4967]: E0120 09:12:56.736306 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558871cd-7ca8-4399-a7d7-346605104bb1" containerName="keystone-api" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.736313 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="558871cd-7ca8-4399-a7d7-346605104bb1" containerName="keystone-api" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.736455 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="558871cd-7ca8-4399-a7d7-346605104bb1" containerName="keystone-api" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.736469 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e009e9d1-fa56-4174-bdd0-29b5e2974b95" containerName="mariadb-account-delete" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.737070 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.740971 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.741965 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.743347 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.745215 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-slwx4"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.758875 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz"] Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.896252 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.896874 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdspp\" (UniqueName: \"kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.896942 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdppg\" (UniqueName: \"kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.896973 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.998316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdppg\" (UniqueName: \"kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.998718 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.998960 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:56 crc kubenswrapper[4967]: I0120 09:12:56.999242 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdspp\" (UniqueName: \"kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:56.999962 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.000041 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.026035 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdspp\" (UniqueName: \"kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp\") pod \"keystone-f510-account-create-update-bkbcz\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.026569 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdppg\" (UniqueName: \"kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg\") pod \"keystone-db-create-slwx4\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.051732 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.065749 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.493201 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-slwx4"] Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.524726 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz"] Jan 20 09:12:57 crc kubenswrapper[4967]: W0120 09:12:57.527367 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f266e18_9998_4fa6_93e5_4a5f3bd09dd7.slice/crio-9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f WatchSource:0}: Error finding container 9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f: Status 404 returned error can't find the container with id 9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.701072 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3287d1c7-fc4d-4449-ad93-369d4631dc72" path="/var/lib/kubelet/pods/3287d1c7-fc4d-4449-ad93-369d4631dc72/volumes" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.701975 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="558871cd-7ca8-4399-a7d7-346605104bb1" path="/var/lib/kubelet/pods/558871cd-7ca8-4399-a7d7-346605104bb1/volumes" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.702623 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3ef008-51cc-4810-89b7-507f156ed524" path="/var/lib/kubelet/pods/dc3ef008-51cc-4810-89b7-507f156ed524/volumes" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.703210 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e009e9d1-fa56-4174-bdd0-29b5e2974b95" path="/var/lib/kubelet/pods/e009e9d1-fa56-4174-bdd0-29b5e2974b95/volumes" Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.917450 4967 generic.go:334] "Generic (PLEG): container finished" podID="0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" containerID="85c48fe3b63a501efb76238b7815533a6da2b2df72897ff06fb2db5198706593" exitCode=0 Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.917513 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" event={"ID":"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7","Type":"ContainerDied","Data":"85c48fe3b63a501efb76238b7815533a6da2b2df72897ff06fb2db5198706593"} Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.917889 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" event={"ID":"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7","Type":"ContainerStarted","Data":"9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f"} Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.919098 4967 generic.go:334] "Generic (PLEG): container finished" podID="7c7bec2c-653c-4349-a178-38f5979917dc" containerID="d475f9b8498897f9c18a47b5ac0b581c2fa8d09b9c1db4f538c2d70b2795c357" exitCode=0 Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.919128 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-slwx4" event={"ID":"7c7bec2c-653c-4349-a178-38f5979917dc","Type":"ContainerDied","Data":"d475f9b8498897f9c18a47b5ac0b581c2fa8d09b9c1db4f538c2d70b2795c357"} Jan 20 09:12:57 crc kubenswrapper[4967]: I0120 09:12:57.919149 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-slwx4" event={"ID":"7c7bec2c-653c-4349-a178-38f5979917dc","Type":"ContainerStarted","Data":"2d78dcceaae5ad1a487f78b6230e9ba3e5b03692c610c6366e9c9eb7a170b2af"} Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.227766 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.233265 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.333326 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts\") pod \"7c7bec2c-653c-4349-a178-38f5979917dc\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.333450 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdppg\" (UniqueName: \"kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg\") pod \"7c7bec2c-653c-4349-a178-38f5979917dc\" (UID: \"7c7bec2c-653c-4349-a178-38f5979917dc\") " Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.334129 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c7bec2c-653c-4349-a178-38f5979917dc" (UID: "7c7bec2c-653c-4349-a178-38f5979917dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.338542 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg" (OuterVolumeSpecName: "kube-api-access-mdppg") pod "7c7bec2c-653c-4349-a178-38f5979917dc" (UID: "7c7bec2c-653c-4349-a178-38f5979917dc"). InnerVolumeSpecName "kube-api-access-mdppg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.434994 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdspp\" (UniqueName: \"kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp\") pod \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.435159 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts\") pod \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\" (UID: \"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7\") " Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.435610 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c7bec2c-653c-4349-a178-38f5979917dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.435643 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdppg\" (UniqueName: \"kubernetes.io/projected/7c7bec2c-653c-4349-a178-38f5979917dc-kube-api-access-mdppg\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.436489 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" (UID: "0f266e18-9998-4fa6-93e5-4a5f3bd09dd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.438495 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp" (OuterVolumeSpecName: "kube-api-access-wdspp") pod "0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" (UID: "0f266e18-9998-4fa6-93e5-4a5f3bd09dd7"). InnerVolumeSpecName "kube-api-access-wdspp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.537911 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdspp\" (UniqueName: \"kubernetes.io/projected/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-kube-api-access-wdspp\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.537949 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.932664 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.932659 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz" event={"ID":"0f266e18-9998-4fa6-93e5-4a5f3bd09dd7","Type":"ContainerDied","Data":"9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f"} Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.932793 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9708913daedab1ac8b47c494d511e77c824479e184bbf9c2575b5d6bb3bc028f" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.934727 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-slwx4" event={"ID":"7c7bec2c-653c-4349-a178-38f5979917dc","Type":"ContainerDied","Data":"2d78dcceaae5ad1a487f78b6230e9ba3e5b03692c610c6366e9c9eb7a170b2af"} Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.934828 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d78dcceaae5ad1a487f78b6230e9ba3e5b03692c610c6366e9c9eb7a170b2af" Jan 20 09:12:59 crc kubenswrapper[4967]: I0120 09:12:59.934979 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-slwx4" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.407471 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-bpct5"] Jan 20 09:13:02 crc kubenswrapper[4967]: E0120 09:13:02.407995 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" containerName="mariadb-account-create-update" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.408013 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" containerName="mariadb-account-create-update" Jan 20 09:13:02 crc kubenswrapper[4967]: E0120 09:13:02.408032 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7bec2c-653c-4349-a178-38f5979917dc" containerName="mariadb-database-create" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.408040 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7bec2c-653c-4349-a178-38f5979917dc" containerName="mariadb-database-create" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.408189 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7bec2c-653c-4349-a178-38f5979917dc" containerName="mariadb-database-create" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.408206 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" containerName="mariadb-account-create-update" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.408809 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.410532 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.410846 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-rpp5h" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.411384 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.411928 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.423720 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-bpct5"] Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.487976 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.488051 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lbsm\" (UniqueName: \"kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.589584 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.589720 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lbsm\" (UniqueName: \"kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.594578 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.612823 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lbsm\" (UniqueName: \"kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm\") pod \"keystone-db-sync-bpct5\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.730941 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.932443 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-bpct5"] Jan 20 09:13:02 crc kubenswrapper[4967]: I0120 09:13:02.954133 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" event={"ID":"c08817a8-fc2d-4504-926d-808d0c224134","Type":"ContainerStarted","Data":"a048c87144154d3212ccd91e1ff4811f52e272142e269a35e16dfa3b820e7fa7"} Jan 20 09:13:03 crc kubenswrapper[4967]: I0120 09:13:03.961728 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" event={"ID":"c08817a8-fc2d-4504-926d-808d0c224134","Type":"ContainerStarted","Data":"47890fa2d12fc82ecec684c9ada7a1689722c0b13db888e00d5757865b74a923"} Jan 20 09:13:03 crc kubenswrapper[4967]: I0120 09:13:03.982291 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" podStartSLOduration=1.98227151 podStartE2EDuration="1.98227151s" podCreationTimestamp="2026-01-20 09:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:13:03.974359382 +0000 UTC m=+1418.559659599" watchObservedRunningTime="2026-01-20 09:13:03.98227151 +0000 UTC m=+1418.567571727" Jan 20 09:13:04 crc kubenswrapper[4967]: I0120 09:13:04.969806 4967 generic.go:334] "Generic (PLEG): container finished" podID="c08817a8-fc2d-4504-926d-808d0c224134" containerID="47890fa2d12fc82ecec684c9ada7a1689722c0b13db888e00d5757865b74a923" exitCode=0 Jan 20 09:13:04 crc kubenswrapper[4967]: I0120 09:13:04.969916 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" event={"ID":"c08817a8-fc2d-4504-926d-808d0c224134","Type":"ContainerDied","Data":"47890fa2d12fc82ecec684c9ada7a1689722c0b13db888e00d5757865b74a923"} Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.210411 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.243116 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lbsm\" (UniqueName: \"kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm\") pod \"c08817a8-fc2d-4504-926d-808d0c224134\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.243187 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data\") pod \"c08817a8-fc2d-4504-926d-808d0c224134\" (UID: \"c08817a8-fc2d-4504-926d-808d0c224134\") " Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.249234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm" (OuterVolumeSpecName: "kube-api-access-2lbsm") pod "c08817a8-fc2d-4504-926d-808d0c224134" (UID: "c08817a8-fc2d-4504-926d-808d0c224134"). InnerVolumeSpecName "kube-api-access-2lbsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.288653 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data" (OuterVolumeSpecName: "config-data") pod "c08817a8-fc2d-4504-926d-808d0c224134" (UID: "c08817a8-fc2d-4504-926d-808d0c224134"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.344591 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lbsm\" (UniqueName: \"kubernetes.io/projected/c08817a8-fc2d-4504-926d-808d0c224134-kube-api-access-2lbsm\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.344641 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c08817a8-fc2d-4504-926d-808d0c224134-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.985274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" event={"ID":"c08817a8-fc2d-4504-926d-808d0c224134","Type":"ContainerDied","Data":"a048c87144154d3212ccd91e1ff4811f52e272142e269a35e16dfa3b820e7fa7"} Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.985570 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a048c87144154d3212ccd91e1ff4811f52e272142e269a35e16dfa3b820e7fa7" Jan 20 09:13:06 crc kubenswrapper[4967]: I0120 09:13:06.985352 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-bpct5" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.166193 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-swbcp"] Jan 20 09:13:07 crc kubenswrapper[4967]: E0120 09:13:07.166520 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08817a8-fc2d-4504-926d-808d0c224134" containerName="keystone-db-sync" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.166535 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08817a8-fc2d-4504-926d-808d0c224134" containerName="keystone-db-sync" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.166715 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08817a8-fc2d-4504-926d-808d0c224134" containerName="keystone-db-sync" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.168300 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.170033 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.170448 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.170671 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-rpp5h" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.170871 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.171126 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.180228 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-swbcp"] Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.257364 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.257404 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.257425 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.257573 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.257746 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrn86\" (UniqueName: \"kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.359740 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.359801 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.359858 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.359914 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrn86\" (UniqueName: \"kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.359995 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.363796 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.363874 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.364004 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.367208 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.381359 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrn86\" (UniqueName: \"kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86\") pod \"keystone-bootstrap-swbcp\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.485412 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.874667 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-swbcp"] Jan 20 09:13:07 crc kubenswrapper[4967]: W0120 09:13:07.878019 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6e211f6_3b38_439e_a6ff_4d256023ac23.slice/crio-f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec WatchSource:0}: Error finding container f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec: Status 404 returned error can't find the container with id f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec Jan 20 09:13:07 crc kubenswrapper[4967]: I0120 09:13:07.999204 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" event={"ID":"b6e211f6-3b38-439e-a6ff-4d256023ac23","Type":"ContainerStarted","Data":"f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec"} Jan 20 09:13:09 crc kubenswrapper[4967]: I0120 09:13:09.008633 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" event={"ID":"b6e211f6-3b38-439e-a6ff-4d256023ac23","Type":"ContainerStarted","Data":"f1256ed292eae1a7d754635ae38ca6e2eb72c33afa8fde8ef55b5ead4073ee5e"} Jan 20 09:13:09 crc kubenswrapper[4967]: I0120 09:13:09.032061 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" podStartSLOduration=2.031218576 podStartE2EDuration="2.031218576s" podCreationTimestamp="2026-01-20 09:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:13:09.025408953 +0000 UTC m=+1423.610709160" watchObservedRunningTime="2026-01-20 09:13:09.031218576 +0000 UTC m=+1423.616518813" Jan 20 09:13:11 crc kubenswrapper[4967]: I0120 09:13:11.025710 4967 generic.go:334] "Generic (PLEG): container finished" podID="b6e211f6-3b38-439e-a6ff-4d256023ac23" containerID="f1256ed292eae1a7d754635ae38ca6e2eb72c33afa8fde8ef55b5ead4073ee5e" exitCode=0 Jan 20 09:13:11 crc kubenswrapper[4967]: I0120 09:13:11.025803 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" event={"ID":"b6e211f6-3b38-439e-a6ff-4d256023ac23","Type":"ContainerDied","Data":"f1256ed292eae1a7d754635ae38ca6e2eb72c33afa8fde8ef55b5ead4073ee5e"} Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.268523 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.348714 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrn86\" (UniqueName: \"kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86\") pod \"b6e211f6-3b38-439e-a6ff-4d256023ac23\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.348756 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts\") pod \"b6e211f6-3b38-439e-a6ff-4d256023ac23\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.348815 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys\") pod \"b6e211f6-3b38-439e-a6ff-4d256023ac23\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.348891 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data\") pod \"b6e211f6-3b38-439e-a6ff-4d256023ac23\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.348956 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys\") pod \"b6e211f6-3b38-439e-a6ff-4d256023ac23\" (UID: \"b6e211f6-3b38-439e-a6ff-4d256023ac23\") " Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.355723 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b6e211f6-3b38-439e-a6ff-4d256023ac23" (UID: "b6e211f6-3b38-439e-a6ff-4d256023ac23"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.356915 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b6e211f6-3b38-439e-a6ff-4d256023ac23" (UID: "b6e211f6-3b38-439e-a6ff-4d256023ac23"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.356963 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts" (OuterVolumeSpecName: "scripts") pod "b6e211f6-3b38-439e-a6ff-4d256023ac23" (UID: "b6e211f6-3b38-439e-a6ff-4d256023ac23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.357223 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86" (OuterVolumeSpecName: "kube-api-access-nrn86") pod "b6e211f6-3b38-439e-a6ff-4d256023ac23" (UID: "b6e211f6-3b38-439e-a6ff-4d256023ac23"). InnerVolumeSpecName "kube-api-access-nrn86". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.370962 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data" (OuterVolumeSpecName: "config-data") pod "b6e211f6-3b38-439e-a6ff-4d256023ac23" (UID: "b6e211f6-3b38-439e-a6ff-4d256023ac23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.450636 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.450920 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrn86\" (UniqueName: \"kubernetes.io/projected/b6e211f6-3b38-439e-a6ff-4d256023ac23-kube-api-access-nrn86\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.450983 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.451036 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:12 crc kubenswrapper[4967]: I0120 09:13:12.451101 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e211f6-3b38-439e-a6ff-4d256023ac23-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.041643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" event={"ID":"b6e211f6-3b38-439e-a6ff-4d256023ac23","Type":"ContainerDied","Data":"f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec"} Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.041685 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0cacbb60fc39d7988a489c82de52cb63af7e828fa6541a2c0ebeeeaee0be6ec" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.041725 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-swbcp" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.123017 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:13:13 crc kubenswrapper[4967]: E0120 09:13:13.123320 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e211f6-3b38-439e-a6ff-4d256023ac23" containerName="keystone-bootstrap" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.123341 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e211f6-3b38-439e-a6ff-4d256023ac23" containerName="keystone-bootstrap" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.123494 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e211f6-3b38-439e-a6ff-4d256023ac23" containerName="keystone-bootstrap" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.124029 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.126106 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.126452 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.126740 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-rpp5h" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.127775 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.154076 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.163391 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.163495 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.163521 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w78h\" (UniqueName: \"kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.163560 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.163582 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.264571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.264612 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w78h\" (UniqueName: \"kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.264665 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.264683 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.264729 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.269461 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.269487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.270239 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.278233 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.283987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w78h\" (UniqueName: \"kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h\") pod \"keystone-b87fd47db-zt4sd\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:13 crc kubenswrapper[4967]: I0120 09:13:13.448277 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:14 crc kubenswrapper[4967]: I0120 09:13:13.843656 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:13:14 crc kubenswrapper[4967]: I0120 09:13:14.054524 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" event={"ID":"00067efd-56e9-4cb4-9878-cc2b98c6400c","Type":"ContainerStarted","Data":"a909c303fbe20b9b59a7eebcc3e8785679f16517433a4138c2647e3d2c1c912f"} Jan 20 09:13:15 crc kubenswrapper[4967]: I0120 09:13:15.068110 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" event={"ID":"00067efd-56e9-4cb4-9878-cc2b98c6400c","Type":"ContainerStarted","Data":"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7"} Jan 20 09:13:15 crc kubenswrapper[4967]: I0120 09:13:15.068477 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:44 crc kubenswrapper[4967]: I0120 09:13:44.904547 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:13:44 crc kubenswrapper[4967]: I0120 09:13:44.930593 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" podStartSLOduration=31.930573299 podStartE2EDuration="31.930573299s" podCreationTimestamp="2026-01-20 09:13:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:13:15.090801743 +0000 UTC m=+1429.676101950" watchObservedRunningTime="2026-01-20 09:13:44.930573299 +0000 UTC m=+1459.515873506" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.683235 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-swbcp"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.692940 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-bpct5"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.699352 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-swbcp"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.735115 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-bpct5"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.745170 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.745546 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" podUID="00067efd-56e9-4cb4-9878-cc2b98c6400c" containerName="keystone-api" containerID="cri-o://4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7" gracePeriod=30 Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.753691 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystonef510-account-delete-ljmdq"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.754553 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.766940 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonef510-account-delete-ljmdq"] Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.844651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.844875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxfwq\" (UniqueName: \"kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.946547 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.946757 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxfwq\" (UniqueName: \"kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.947729 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:02 crc kubenswrapper[4967]: I0120 09:14:02.972230 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxfwq\" (UniqueName: \"kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq\") pod \"keystonef510-account-delete-ljmdq\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:03 crc kubenswrapper[4967]: I0120 09:14:03.075751 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:03 crc kubenswrapper[4967]: I0120 09:14:03.464404 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonef510-account-delete-ljmdq"] Jan 20 09:14:03 crc kubenswrapper[4967]: I0120 09:14:03.702296 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e211f6-3b38-439e-a6ff-4d256023ac23" path="/var/lib/kubelet/pods/b6e211f6-3b38-439e-a6ff-4d256023ac23/volumes" Jan 20 09:14:03 crc kubenswrapper[4967]: I0120 09:14:03.703298 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c08817a8-fc2d-4504-926d-808d0c224134" path="/var/lib/kubelet/pods/c08817a8-fc2d-4504-926d-808d0c224134/volumes" Jan 20 09:14:04 crc kubenswrapper[4967]: I0120 09:14:04.407521 4967 generic.go:334] "Generic (PLEG): container finished" podID="92115904-4466-47ce-9114-da5fcff908b2" containerID="e9efeeafcd07e757fe67dc9b21d531b2b08b8097d021a9dd6b63c1ee0dea12f2" exitCode=0 Jan 20 09:14:04 crc kubenswrapper[4967]: I0120 09:14:04.407590 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" event={"ID":"92115904-4466-47ce-9114-da5fcff908b2","Type":"ContainerDied","Data":"e9efeeafcd07e757fe67dc9b21d531b2b08b8097d021a9dd6b63c1ee0dea12f2"} Jan 20 09:14:04 crc kubenswrapper[4967]: I0120 09:14:04.408327 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" event={"ID":"92115904-4466-47ce-9114-da5fcff908b2","Type":"ContainerStarted","Data":"707fd5129ebfdf61ef95be02503408e77715ca8e822bb702c863e7b36fad6f01"} Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.747982 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.886663 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts\") pod \"92115904-4466-47ce-9114-da5fcff908b2\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.886797 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxfwq\" (UniqueName: \"kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq\") pod \"92115904-4466-47ce-9114-da5fcff908b2\" (UID: \"92115904-4466-47ce-9114-da5fcff908b2\") " Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.887599 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92115904-4466-47ce-9114-da5fcff908b2" (UID: "92115904-4466-47ce-9114-da5fcff908b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.893376 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq" (OuterVolumeSpecName: "kube-api-access-qxfwq") pod "92115904-4466-47ce-9114-da5fcff908b2" (UID: "92115904-4466-47ce-9114-da5fcff908b2"). InnerVolumeSpecName "kube-api-access-qxfwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.987857 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxfwq\" (UniqueName: \"kubernetes.io/projected/92115904-4466-47ce-9114-da5fcff908b2-kube-api-access-qxfwq\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:05 crc kubenswrapper[4967]: I0120 09:14:05.987885 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92115904-4466-47ce-9114-da5fcff908b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.155760 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.290410 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys\") pod \"00067efd-56e9-4cb4-9878-cc2b98c6400c\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.290649 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts\") pod \"00067efd-56e9-4cb4-9878-cc2b98c6400c\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.290742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys\") pod \"00067efd-56e9-4cb4-9878-cc2b98c6400c\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.290779 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data\") pod \"00067efd-56e9-4cb4-9878-cc2b98c6400c\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.290802 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w78h\" (UniqueName: \"kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h\") pod \"00067efd-56e9-4cb4-9878-cc2b98c6400c\" (UID: \"00067efd-56e9-4cb4-9878-cc2b98c6400c\") " Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.293503 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "00067efd-56e9-4cb4-9878-cc2b98c6400c" (UID: "00067efd-56e9-4cb4-9878-cc2b98c6400c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.293542 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts" (OuterVolumeSpecName: "scripts") pod "00067efd-56e9-4cb4-9878-cc2b98c6400c" (UID: "00067efd-56e9-4cb4-9878-cc2b98c6400c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.294527 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "00067efd-56e9-4cb4-9878-cc2b98c6400c" (UID: "00067efd-56e9-4cb4-9878-cc2b98c6400c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.294782 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h" (OuterVolumeSpecName: "kube-api-access-4w78h") pod "00067efd-56e9-4cb4-9878-cc2b98c6400c" (UID: "00067efd-56e9-4cb4-9878-cc2b98c6400c"). InnerVolumeSpecName "kube-api-access-4w78h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.315274 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data" (OuterVolumeSpecName: "config-data") pod "00067efd-56e9-4cb4-9878-cc2b98c6400c" (UID: "00067efd-56e9-4cb4-9878-cc2b98c6400c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.391944 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.391975 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.391985 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w78h\" (UniqueName: \"kubernetes.io/projected/00067efd-56e9-4cb4-9878-cc2b98c6400c-kube-api-access-4w78h\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.391997 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.392005 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00067efd-56e9-4cb4-9878-cc2b98c6400c-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.423343 4967 generic.go:334] "Generic (PLEG): container finished" podID="00067efd-56e9-4cb4-9878-cc2b98c6400c" containerID="4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7" exitCode=0 Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.423465 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.423867 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" event={"ID":"00067efd-56e9-4cb4-9878-cc2b98c6400c","Type":"ContainerDied","Data":"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7"} Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.423912 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-b87fd47db-zt4sd" event={"ID":"00067efd-56e9-4cb4-9878-cc2b98c6400c","Type":"ContainerDied","Data":"a909c303fbe20b9b59a7eebcc3e8785679f16517433a4138c2647e3d2c1c912f"} Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.423933 4967 scope.go:117] "RemoveContainer" containerID="4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.426354 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" event={"ID":"92115904-4466-47ce-9114-da5fcff908b2","Type":"ContainerDied","Data":"707fd5129ebfdf61ef95be02503408e77715ca8e822bb702c863e7b36fad6f01"} Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.426460 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="707fd5129ebfdf61ef95be02503408e77715ca8e822bb702c863e7b36fad6f01" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.426540 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonef510-account-delete-ljmdq" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.452379 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.454095 4967 scope.go:117] "RemoveContainer" containerID="4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7" Jan 20 09:14:06 crc kubenswrapper[4967]: E0120 09:14:06.454420 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7\": container with ID starting with 4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7 not found: ID does not exist" containerID="4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.454447 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7"} err="failed to get container status \"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7\": rpc error: code = NotFound desc = could not find container \"4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7\": container with ID starting with 4b671f81863887efda4a0cf57da7d55f89bdbdc6adf0ebb61d106ff89a0ccdc7 not found: ID does not exist" Jan 20 09:14:06 crc kubenswrapper[4967]: I0120 09:14:06.456898 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-b87fd47db-zt4sd"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.704298 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00067efd-56e9-4cb4-9878-cc2b98c6400c" path="/var/lib/kubelet/pods/00067efd-56e9-4cb4-9878-cc2b98c6400c/volumes" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.748903 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-slwx4"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.756019 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-slwx4"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.766484 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.771328 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystonef510-account-delete-ljmdq"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.776958 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-f510-account-create-update-bkbcz"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.782350 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystonef510-account-delete-ljmdq"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.841715 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-create-qv4qt"] Jan 20 09:14:07 crc kubenswrapper[4967]: E0120 09:14:07.841950 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92115904-4466-47ce-9114-da5fcff908b2" containerName="mariadb-account-delete" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.841962 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92115904-4466-47ce-9114-da5fcff908b2" containerName="mariadb-account-delete" Jan 20 09:14:07 crc kubenswrapper[4967]: E0120 09:14:07.841976 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00067efd-56e9-4cb4-9878-cc2b98c6400c" containerName="keystone-api" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.841983 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="00067efd-56e9-4cb4-9878-cc2b98c6400c" containerName="keystone-api" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.842081 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="92115904-4466-47ce-9114-da5fcff908b2" containerName="mariadb-account-delete" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.842105 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="00067efd-56e9-4cb4-9878-cc2b98c6400c" containerName="keystone-api" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.842582 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.862296 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-qv4qt"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.944635 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-c611-account-create-update-glt4n"] Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.945535 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.947439 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-db-secret" Jan 20 09:14:07 crc kubenswrapper[4967]: I0120 09:14:07.958088 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-c611-account-create-update-glt4n"] Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.014715 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlf5m\" (UniqueName: \"kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.014773 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.116177 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkw2m\" (UniqueName: \"kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.116280 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.116388 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlf5m\" (UniqueName: \"kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.116488 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.117251 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.140481 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlf5m\" (UniqueName: \"kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m\") pod \"keystone-db-create-qv4qt\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.158127 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.217766 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkw2m\" (UniqueName: \"kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.217850 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.218645 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.259209 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkw2m\" (UniqueName: \"kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m\") pod \"keystone-c611-account-create-update-glt4n\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.557805 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.578381 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-qv4qt"] Jan 20 09:14:08 crc kubenswrapper[4967]: I0120 09:14:08.951237 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-c611-account-create-update-glt4n"] Jan 20 09:14:08 crc kubenswrapper[4967]: W0120 09:14:08.958962 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99758e56_9a70_42da_af60_10cd9ffcacad.slice/crio-7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67 WatchSource:0}: Error finding container 7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67: Status 404 returned error can't find the container with id 7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67 Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.454140 4967 generic.go:334] "Generic (PLEG): container finished" podID="1bb7901f-4270-4905-a23b-1dba2d205315" containerID="4ea689945fbcce99f3caa54beeceab17153ccef9a71af7cce581407b4be2b0c4" exitCode=0 Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.454274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" event={"ID":"1bb7901f-4270-4905-a23b-1dba2d205315","Type":"ContainerDied","Data":"4ea689945fbcce99f3caa54beeceab17153ccef9a71af7cce581407b4be2b0c4"} Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.454532 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" event={"ID":"1bb7901f-4270-4905-a23b-1dba2d205315","Type":"ContainerStarted","Data":"4005dd02620d3f08d80a936c9cfd1b3e76558cb220a98cc5f03e535c1ea81f3c"} Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.458641 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" event={"ID":"99758e56-9a70-42da-af60-10cd9ffcacad","Type":"ContainerDied","Data":"c9ebcb9214c544fcaf6c7da0c9092d0365c183dcbd8d19a109313de62197d4b0"} Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.458602 4967 generic.go:334] "Generic (PLEG): container finished" podID="99758e56-9a70-42da-af60-10cd9ffcacad" containerID="c9ebcb9214c544fcaf6c7da0c9092d0365c183dcbd8d19a109313de62197d4b0" exitCode=0 Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.458724 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" event={"ID":"99758e56-9a70-42da-af60-10cd9ffcacad","Type":"ContainerStarted","Data":"7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67"} Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.702012 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f266e18-9998-4fa6-93e5-4a5f3bd09dd7" path="/var/lib/kubelet/pods/0f266e18-9998-4fa6-93e5-4a5f3bd09dd7/volumes" Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.702610 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c7bec2c-653c-4349-a178-38f5979917dc" path="/var/lib/kubelet/pods/7c7bec2c-653c-4349-a178-38f5979917dc/volumes" Jan 20 09:14:09 crc kubenswrapper[4967]: I0120 09:14:09.703262 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92115904-4466-47ce-9114-da5fcff908b2" path="/var/lib/kubelet/pods/92115904-4466-47ce-9114-da5fcff908b2/volumes" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.744434 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.819212 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.855878 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts\") pod \"99758e56-9a70-42da-af60-10cd9ffcacad\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.855991 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkw2m\" (UniqueName: \"kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m\") pod \"99758e56-9a70-42da-af60-10cd9ffcacad\" (UID: \"99758e56-9a70-42da-af60-10cd9ffcacad\") " Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.856456 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99758e56-9a70-42da-af60-10cd9ffcacad" (UID: "99758e56-9a70-42da-af60-10cd9ffcacad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.862194 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m" (OuterVolumeSpecName: "kube-api-access-fkw2m") pod "99758e56-9a70-42da-af60-10cd9ffcacad" (UID: "99758e56-9a70-42da-af60-10cd9ffcacad"). InnerVolumeSpecName "kube-api-access-fkw2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.957009 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts\") pod \"1bb7901f-4270-4905-a23b-1dba2d205315\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.957067 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlf5m\" (UniqueName: \"kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m\") pod \"1bb7901f-4270-4905-a23b-1dba2d205315\" (UID: \"1bb7901f-4270-4905-a23b-1dba2d205315\") " Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.957410 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99758e56-9a70-42da-af60-10cd9ffcacad-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.957430 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkw2m\" (UniqueName: \"kubernetes.io/projected/99758e56-9a70-42da-af60-10cd9ffcacad-kube-api-access-fkw2m\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.957661 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1bb7901f-4270-4905-a23b-1dba2d205315" (UID: "1bb7901f-4270-4905-a23b-1dba2d205315"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:14:10 crc kubenswrapper[4967]: I0120 09:14:10.959757 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m" (OuterVolumeSpecName: "kube-api-access-dlf5m") pod "1bb7901f-4270-4905-a23b-1dba2d205315" (UID: "1bb7901f-4270-4905-a23b-1dba2d205315"). InnerVolumeSpecName "kube-api-access-dlf5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.058868 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1bb7901f-4270-4905-a23b-1dba2d205315-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.058917 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlf5m\" (UniqueName: \"kubernetes.io/projected/1bb7901f-4270-4905-a23b-1dba2d205315-kube-api-access-dlf5m\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.477413 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" event={"ID":"99758e56-9a70-42da-af60-10cd9ffcacad","Type":"ContainerDied","Data":"7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67"} Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.477501 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b92ae3ae2980e8bf45eefe6c23530c33bfb878af3887836958e779a9c2ffa67" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.477450 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-c611-account-create-update-glt4n" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.479731 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" event={"ID":"1bb7901f-4270-4905-a23b-1dba2d205315","Type":"ContainerDied","Data":"4005dd02620d3f08d80a936c9cfd1b3e76558cb220a98cc5f03e535c1ea81f3c"} Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.479763 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-create-qv4qt" Jan 20 09:14:11 crc kubenswrapper[4967]: I0120 09:14:11.479775 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4005dd02620d3f08d80a936c9cfd1b3e76558cb220a98cc5f03e535c1ea81f3c" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.511296 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f6bfl"] Jan 20 09:14:13 crc kubenswrapper[4967]: E0120 09:14:13.511912 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb7901f-4270-4905-a23b-1dba2d205315" containerName="mariadb-database-create" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.511930 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb7901f-4270-4905-a23b-1dba2d205315" containerName="mariadb-database-create" Jan 20 09:14:13 crc kubenswrapper[4967]: E0120 09:14:13.511950 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99758e56-9a70-42da-af60-10cd9ffcacad" containerName="mariadb-account-create-update" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.511959 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="99758e56-9a70-42da-af60-10cd9ffcacad" containerName="mariadb-account-create-update" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.512107 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb7901f-4270-4905-a23b-1dba2d205315" containerName="mariadb-database-create" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.512123 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="99758e56-9a70-42da-af60-10cd9ffcacad" containerName="mariadb-account-create-update" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.512672 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.514922 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.517950 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.518698 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7g8vh" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.518715 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.526882 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f6bfl"] Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.697697 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.697837 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95lcj\" (UniqueName: \"kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.799420 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95lcj\" (UniqueName: \"kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.799507 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.810546 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.820064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95lcj\" (UniqueName: \"kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj\") pod \"keystone-db-sync-f6bfl\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:13 crc kubenswrapper[4967]: I0120 09:14:13.837714 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:14 crc kubenswrapper[4967]: I0120 09:14:14.257086 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f6bfl"] Jan 20 09:14:14 crc kubenswrapper[4967]: I0120 09:14:14.503377 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" event={"ID":"ee3c6996-85b2-42d6-879b-f4c3acbe6e75","Type":"ContainerStarted","Data":"e4060a98764541a9fe8ea245229b415f396c644ccb527a93f49ec8debf047923"} Jan 20 09:14:14 crc kubenswrapper[4967]: I0120 09:14:14.503744 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" event={"ID":"ee3c6996-85b2-42d6-879b-f4c3acbe6e75","Type":"ContainerStarted","Data":"363e292a592ae2b3edd006f296f49cb3b23c1d2c2efbe9c4e65c9e6fcc137eac"} Jan 20 09:14:16 crc kubenswrapper[4967]: I0120 09:14:16.518571 4967 generic.go:334] "Generic (PLEG): container finished" podID="ee3c6996-85b2-42d6-879b-f4c3acbe6e75" containerID="e4060a98764541a9fe8ea245229b415f396c644ccb527a93f49ec8debf047923" exitCode=0 Jan 20 09:14:16 crc kubenswrapper[4967]: I0120 09:14:16.518635 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" event={"ID":"ee3c6996-85b2-42d6-879b-f4c3acbe6e75","Type":"ContainerDied","Data":"e4060a98764541a9fe8ea245229b415f396c644ccb527a93f49ec8debf047923"} Jan 20 09:14:17 crc kubenswrapper[4967]: I0120 09:14:17.788901 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:17 crc kubenswrapper[4967]: I0120 09:14:17.953308 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data\") pod \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " Jan 20 09:14:17 crc kubenswrapper[4967]: I0120 09:14:17.953401 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95lcj\" (UniqueName: \"kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj\") pod \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\" (UID: \"ee3c6996-85b2-42d6-879b-f4c3acbe6e75\") " Jan 20 09:14:17 crc kubenswrapper[4967]: I0120 09:14:17.959754 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj" (OuterVolumeSpecName: "kube-api-access-95lcj") pod "ee3c6996-85b2-42d6-879b-f4c3acbe6e75" (UID: "ee3c6996-85b2-42d6-879b-f4c3acbe6e75"). InnerVolumeSpecName "kube-api-access-95lcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:17 crc kubenswrapper[4967]: I0120 09:14:17.996808 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data" (OuterVolumeSpecName: "config-data") pod "ee3c6996-85b2-42d6-879b-f4c3acbe6e75" (UID: "ee3c6996-85b2-42d6-879b-f4c3acbe6e75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.054592 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.054650 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95lcj\" (UniqueName: \"kubernetes.io/projected/ee3c6996-85b2-42d6-879b-f4c3acbe6e75-kube-api-access-95lcj\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.535656 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" event={"ID":"ee3c6996-85b2-42d6-879b-f4c3acbe6e75","Type":"ContainerDied","Data":"363e292a592ae2b3edd006f296f49cb3b23c1d2c2efbe9c4e65c9e6fcc137eac"} Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.535717 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="363e292a592ae2b3edd006f296f49cb3b23c1d2c2efbe9c4e65c9e6fcc137eac" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.535849 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-db-sync-f6bfl" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.677654 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-755cf"] Jan 20 09:14:18 crc kubenswrapper[4967]: E0120 09:14:18.678041 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3c6996-85b2-42d6-879b-f4c3acbe6e75" containerName="keystone-db-sync" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.678067 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3c6996-85b2-42d6-879b-f4c3acbe6e75" containerName="keystone-db-sync" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.678252 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3c6996-85b2-42d6-879b-f4c3acbe6e75" containerName="keystone-db-sync" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.678959 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.682239 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.684181 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"osp-secret" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.684228 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.684343 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7g8vh" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.687680 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-755cf"] Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.688425 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.764749 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.765235 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.765329 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.765473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.765562 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7jhw\" (UniqueName: \"kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.866884 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.867421 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.868043 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7jhw\" (UniqueName: \"kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.868675 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.868827 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.871178 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.871563 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.871774 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.873628 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.883489 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7jhw\" (UniqueName: \"kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw\") pod \"keystone-bootstrap-755cf\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:18 crc kubenswrapper[4967]: I0120 09:14:18.996392 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:19 crc kubenswrapper[4967]: I0120 09:14:19.385464 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-755cf"] Jan 20 09:14:19 crc kubenswrapper[4967]: I0120 09:14:19.544414 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" event={"ID":"c8d43314-9260-4e86-9003-6bd022c9c390","Type":"ContainerStarted","Data":"1ca5ddce9967292ff791497f935aa85cbb11da5ea7cfdddff83b505f92848cd1"} Jan 20 09:14:20 crc kubenswrapper[4967]: I0120 09:14:20.551910 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" event={"ID":"c8d43314-9260-4e86-9003-6bd022c9c390","Type":"ContainerStarted","Data":"dede0cf90ddacbad57128ccc6624702d6f2d87714b4d5d531248a9f2c7459ffa"} Jan 20 09:14:20 crc kubenswrapper[4967]: I0120 09:14:20.569967 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" podStartSLOduration=2.5699496760000002 podStartE2EDuration="2.569949676s" podCreationTimestamp="2026-01-20 09:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:14:20.566757729 +0000 UTC m=+1495.152057936" watchObservedRunningTime="2026-01-20 09:14:20.569949676 +0000 UTC m=+1495.155249883" Jan 20 09:14:23 crc kubenswrapper[4967]: I0120 09:14:23.573350 4967 generic.go:334] "Generic (PLEG): container finished" podID="c8d43314-9260-4e86-9003-6bd022c9c390" containerID="dede0cf90ddacbad57128ccc6624702d6f2d87714b4d5d531248a9f2c7459ffa" exitCode=0 Jan 20 09:14:23 crc kubenswrapper[4967]: I0120 09:14:23.573468 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" event={"ID":"c8d43314-9260-4e86-9003-6bd022c9c390","Type":"ContainerDied","Data":"dede0cf90ddacbad57128ccc6624702d6f2d87714b4d5d531248a9f2c7459ffa"} Jan 20 09:14:24 crc kubenswrapper[4967]: I0120 09:14:24.882516 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.059458 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data\") pod \"c8d43314-9260-4e86-9003-6bd022c9c390\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.059554 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys\") pod \"c8d43314-9260-4e86-9003-6bd022c9c390\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.059602 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys\") pod \"c8d43314-9260-4e86-9003-6bd022c9c390\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.059678 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts\") pod \"c8d43314-9260-4e86-9003-6bd022c9c390\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.059734 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7jhw\" (UniqueName: \"kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw\") pod \"c8d43314-9260-4e86-9003-6bd022c9c390\" (UID: \"c8d43314-9260-4e86-9003-6bd022c9c390\") " Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.070831 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c8d43314-9260-4e86-9003-6bd022c9c390" (UID: "c8d43314-9260-4e86-9003-6bd022c9c390"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.070847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c8d43314-9260-4e86-9003-6bd022c9c390" (UID: "c8d43314-9260-4e86-9003-6bd022c9c390"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.070901 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts" (OuterVolumeSpecName: "scripts") pod "c8d43314-9260-4e86-9003-6bd022c9c390" (UID: "c8d43314-9260-4e86-9003-6bd022c9c390"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.072706 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw" (OuterVolumeSpecName: "kube-api-access-z7jhw") pod "c8d43314-9260-4e86-9003-6bd022c9c390" (UID: "c8d43314-9260-4e86-9003-6bd022c9c390"). InnerVolumeSpecName "kube-api-access-z7jhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.078407 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data" (OuterVolumeSpecName: "config-data") pod "c8d43314-9260-4e86-9003-6bd022c9c390" (UID: "c8d43314-9260-4e86-9003-6bd022c9c390"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.161909 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.161966 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7jhw\" (UniqueName: \"kubernetes.io/projected/c8d43314-9260-4e86-9003-6bd022c9c390-kube-api-access-z7jhw\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.161980 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.161992 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.162003 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c8d43314-9260-4e86-9003-6bd022c9c390-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.593276 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" event={"ID":"c8d43314-9260-4e86-9003-6bd022c9c390","Type":"ContainerDied","Data":"1ca5ddce9967292ff791497f935aa85cbb11da5ea7cfdddff83b505f92848cd1"} Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.593663 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ca5ddce9967292ff791497f935aa85cbb11da5ea7cfdddff83b505f92848cd1" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.593331 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-bootstrap-755cf" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.664104 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:14:25 crc kubenswrapper[4967]: E0120 09:14:25.664395 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d43314-9260-4e86-9003-6bd022c9c390" containerName="keystone-bootstrap" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.664408 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d43314-9260-4e86-9003-6bd022c9c390" containerName="keystone-bootstrap" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.664541 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d43314-9260-4e86-9003-6bd022c9c390" containerName="keystone-bootstrap" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.665664 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.667622 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-config-data" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.668075 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-keystone-dockercfg-7g8vh" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.668244 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.668409 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"keystone-scripts" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.675737 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.770657 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pvzh\" (UniqueName: \"kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.770967 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.771123 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.771218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.771348 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.872437 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.872483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.872528 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.872599 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pvzh\" (UniqueName: \"kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.872654 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.877024 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.877057 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.877152 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.878654 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:25 crc kubenswrapper[4967]: I0120 09:14:25.891628 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pvzh\" (UniqueName: \"kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh\") pod \"keystone-7d48b4856d-x6ktr\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:26 crc kubenswrapper[4967]: I0120 09:14:26.003311 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:26 crc kubenswrapper[4967]: I0120 09:14:26.480218 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:14:26 crc kubenswrapper[4967]: I0120 09:14:26.600989 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" event={"ID":"8e0e5e9b-0600-403a-a4b6-ecb254da3396","Type":"ContainerStarted","Data":"8efb0f9f0b899016b17879e45ec51300c3fcd4c2f95a390e34fe7e593ee76aed"} Jan 20 09:14:27 crc kubenswrapper[4967]: I0120 09:14:27.611036 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" event={"ID":"8e0e5e9b-0600-403a-a4b6-ecb254da3396","Type":"ContainerStarted","Data":"278537679b0b42a5c14d897c47331896aed384ed89feda18d3ccd00bd94f9416"} Jan 20 09:14:27 crc kubenswrapper[4967]: I0120 09:14:27.611336 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:27 crc kubenswrapper[4967]: I0120 09:14:27.632647 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" podStartSLOduration=2.632606532 podStartE2EDuration="2.632606532s" podCreationTimestamp="2026-01-20 09:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:14:27.629645891 +0000 UTC m=+1502.214946128" watchObservedRunningTime="2026-01-20 09:14:27.632606532 +0000 UTC m=+1502.217906749" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.160569 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.163860 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.171665 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.286409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.286525 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.286557 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzxrf\" (UniqueName: \"kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.387939 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.388207 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.388289 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzxrf\" (UniqueName: \"kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.388563 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.388893 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.413869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzxrf\" (UniqueName: \"kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf\") pod \"community-operators-qkzw9\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:43 crc kubenswrapper[4967]: I0120 09:14:43.485150 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:44 crc kubenswrapper[4967]: I0120 09:14:44.145988 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:44 crc kubenswrapper[4967]: I0120 09:14:44.758826 4967 generic.go:334] "Generic (PLEG): container finished" podID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerID="37434b546964122d1030a768aab21f29aeb373e806974b80581baefd676eca79" exitCode=0 Jan 20 09:14:44 crc kubenswrapper[4967]: I0120 09:14:44.758879 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerDied","Data":"37434b546964122d1030a768aab21f29aeb373e806974b80581baefd676eca79"} Jan 20 09:14:44 crc kubenswrapper[4967]: I0120 09:14:44.758957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerStarted","Data":"0fe1a9d891b846f2ae526cb1867470a7e08ed669f21a699c85597400c46b9b00"} Jan 20 09:14:44 crc kubenswrapper[4967]: I0120 09:14:44.760896 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 09:14:45 crc kubenswrapper[4967]: I0120 09:14:45.768715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerStarted","Data":"c126592afe33c35a550fd83bac0ee1fea8427db6dfe8317204e975d179779af3"} Jan 20 09:14:46 crc kubenswrapper[4967]: I0120 09:14:46.779208 4967 generic.go:334] "Generic (PLEG): container finished" podID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerID="c126592afe33c35a550fd83bac0ee1fea8427db6dfe8317204e975d179779af3" exitCode=0 Jan 20 09:14:46 crc kubenswrapper[4967]: I0120 09:14:46.779262 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerDied","Data":"c126592afe33c35a550fd83bac0ee1fea8427db6dfe8317204e975d179779af3"} Jan 20 09:14:48 crc kubenswrapper[4967]: I0120 09:14:48.474997 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:14:48 crc kubenswrapper[4967]: I0120 09:14:48.475366 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:14:48 crc kubenswrapper[4967]: I0120 09:14:48.796217 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerStarted","Data":"b032c869a0272f26ffe5c9c1138bc16b369575a8294ad22e6a3151b21e30561c"} Jan 20 09:14:48 crc kubenswrapper[4967]: I0120 09:14:48.817137 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qkzw9" podStartSLOduration=3.227974822 podStartE2EDuration="5.817106645s" podCreationTimestamp="2026-01-20 09:14:43 +0000 UTC" firstStartedPulling="2026-01-20 09:14:44.760675483 +0000 UTC m=+1519.345975690" lastFinishedPulling="2026-01-20 09:14:47.349807306 +0000 UTC m=+1521.935107513" observedRunningTime="2026-01-20 09:14:48.810585366 +0000 UTC m=+1523.395885583" watchObservedRunningTime="2026-01-20 09:14:48.817106645 +0000 UTC m=+1523.402406852" Jan 20 09:14:53 crc kubenswrapper[4967]: I0120 09:14:53.486006 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:53 crc kubenswrapper[4967]: I0120 09:14:53.488020 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:53 crc kubenswrapper[4967]: I0120 09:14:53.548516 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:53 crc kubenswrapper[4967]: I0120 09:14:53.871378 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:53 crc kubenswrapper[4967]: I0120 09:14:53.915586 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:55 crc kubenswrapper[4967]: I0120 09:14:55.855278 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qkzw9" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="registry-server" containerID="cri-o://b032c869a0272f26ffe5c9c1138bc16b369575a8294ad22e6a3151b21e30561c" gracePeriod=2 Jan 20 09:14:57 crc kubenswrapper[4967]: I0120 09:14:57.398167 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:14:57 crc kubenswrapper[4967]: I0120 09:14:57.874804 4967 generic.go:334] "Generic (PLEG): container finished" podID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerID="b032c869a0272f26ffe5c9c1138bc16b369575a8294ad22e6a3151b21e30561c" exitCode=0 Jan 20 09:14:57 crc kubenswrapper[4967]: I0120 09:14:57.874848 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerDied","Data":"b032c869a0272f26ffe5c9c1138bc16b369575a8294ad22e6a3151b21e30561c"} Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.153569 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.227842 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities\") pod \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.227887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content\") pod \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.227941 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzxrf\" (UniqueName: \"kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf\") pod \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\" (UID: \"9a7298f7-8b15-40ee-b7e3-b17f4e647db8\") " Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.229642 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities" (OuterVolumeSpecName: "utilities") pod "9a7298f7-8b15-40ee-b7e3-b17f4e647db8" (UID: "9a7298f7-8b15-40ee-b7e3-b17f4e647db8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.232871 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf" (OuterVolumeSpecName: "kube-api-access-vzxrf") pod "9a7298f7-8b15-40ee-b7e3-b17f4e647db8" (UID: "9a7298f7-8b15-40ee-b7e3-b17f4e647db8"). InnerVolumeSpecName "kube-api-access-vzxrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.274450 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a7298f7-8b15-40ee-b7e3-b17f4e647db8" (UID: "9a7298f7-8b15-40ee-b7e3-b17f4e647db8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.329566 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.329601 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.329649 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzxrf\" (UniqueName: \"kubernetes.io/projected/9a7298f7-8b15-40ee-b7e3-b17f4e647db8-kube-api-access-vzxrf\") on node \"crc\" DevicePath \"\"" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.659744 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:14:58 crc kubenswrapper[4967]: E0120 09:14:58.660085 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="registry-server" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.660103 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="registry-server" Jan 20 09:14:58 crc kubenswrapper[4967]: E0120 09:14:58.660127 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="extract-utilities" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.660136 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="extract-utilities" Jan 20 09:14:58 crc kubenswrapper[4967]: E0120 09:14:58.660144 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="extract-content" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.660153 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="extract-content" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.660298 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" containerName="registry-server" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.660739 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.662657 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"openstack-config-secret" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.662911 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"keystone-kuttl-tests"/"openstack-config" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.665038 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"default-dockercfg-pgbz7" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.668064 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.734817 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.735119 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.735140 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5mx8\" (UniqueName: \"kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.836075 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.836147 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.836164 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5mx8\" (UniqueName: \"kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.837496 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.844124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.866519 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5mx8\" (UniqueName: \"kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8\") pod \"openstackclient\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.883969 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qkzw9" event={"ID":"9a7298f7-8b15-40ee-b7e3-b17f4e647db8","Type":"ContainerDied","Data":"0fe1a9d891b846f2ae526cb1867470a7e08ed669f21a699c85597400c46b9b00"} Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.884017 4967 scope.go:117] "RemoveContainer" containerID="b032c869a0272f26ffe5c9c1138bc16b369575a8294ad22e6a3151b21e30561c" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.884042 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qkzw9" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.907519 4967 scope.go:117] "RemoveContainer" containerID="c126592afe33c35a550fd83bac0ee1fea8427db6dfe8317204e975d179779af3" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.919231 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.929798 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qkzw9"] Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.943450 4967 scope.go:117] "RemoveContainer" containerID="37434b546964122d1030a768aab21f29aeb373e806974b80581baefd676eca79" Jan 20 09:14:58 crc kubenswrapper[4967]: I0120 09:14:58.977143 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Jan 20 09:14:59 crc kubenswrapper[4967]: I0120 09:14:59.371859 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:14:59 crc kubenswrapper[4967]: W0120 09:14:59.374115 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod556459c1_0698_48cb_a63e_125c3f2b2641.slice/crio-300511ee8d0b5f0c14b57058580fa7bfc99df23200afbfc27b62822f38117869 WatchSource:0}: Error finding container 300511ee8d0b5f0c14b57058580fa7bfc99df23200afbfc27b62822f38117869: Status 404 returned error can't find the container with id 300511ee8d0b5f0c14b57058580fa7bfc99df23200afbfc27b62822f38117869 Jan 20 09:14:59 crc kubenswrapper[4967]: I0120 09:14:59.701072 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a7298f7-8b15-40ee-b7e3-b17f4e647db8" path="/var/lib/kubelet/pods/9a7298f7-8b15-40ee-b7e3-b17f4e647db8/volumes" Jan 20 09:14:59 crc kubenswrapper[4967]: I0120 09:14:59.893510 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"556459c1-0698-48cb-a63e-125c3f2b2641","Type":"ContainerStarted","Data":"300511ee8d0b5f0c14b57058580fa7bfc99df23200afbfc27b62822f38117869"} Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.135938 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd"] Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.136840 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.140347 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.145978 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.145980 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd"] Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.154857 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.154932 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.154973 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp7nr\" (UniqueName: \"kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.256834 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp7nr\" (UniqueName: \"kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.257045 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.257098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.258670 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.264236 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.281647 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp7nr\" (UniqueName: \"kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr\") pod \"collect-profiles-29481675-bj9rd\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.458536 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.681563 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd"] Jan 20 09:15:00 crc kubenswrapper[4967]: W0120 09:15:00.692582 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fddd336_9fea_4537_8aec_7c7eb3654949.slice/crio-c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e WatchSource:0}: Error finding container c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e: Status 404 returned error can't find the container with id c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.902120 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" event={"ID":"7fddd336-9fea-4537-8aec-7c7eb3654949","Type":"ContainerStarted","Data":"7a9f4e01de976c62bd20bd31acab47d09cf044ed39a9471e637b2b1e928eae8f"} Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.902157 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" event={"ID":"7fddd336-9fea-4537-8aec-7c7eb3654949","Type":"ContainerStarted","Data":"c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e"} Jan 20 09:15:00 crc kubenswrapper[4967]: I0120 09:15:00.922215 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" podStartSLOduration=0.922198131 podStartE2EDuration="922.198131ms" podCreationTimestamp="2026-01-20 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:15:00.917267555 +0000 UTC m=+1535.502567762" watchObservedRunningTime="2026-01-20 09:15:00.922198131 +0000 UTC m=+1535.507498338" Jan 20 09:15:01 crc kubenswrapper[4967]: I0120 09:15:01.910926 4967 generic.go:334] "Generic (PLEG): container finished" podID="7fddd336-9fea-4537-8aec-7c7eb3654949" containerID="7a9f4e01de976c62bd20bd31acab47d09cf044ed39a9471e637b2b1e928eae8f" exitCode=0 Jan 20 09:15:01 crc kubenswrapper[4967]: I0120 09:15:01.911171 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" event={"ID":"7fddd336-9fea-4537-8aec-7c7eb3654949","Type":"ContainerDied","Data":"7a9f4e01de976c62bd20bd31acab47d09cf044ed39a9471e637b2b1e928eae8f"} Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.608315 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.757011 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp7nr\" (UniqueName: \"kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr\") pod \"7fddd336-9fea-4537-8aec-7c7eb3654949\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.757123 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume\") pod \"7fddd336-9fea-4537-8aec-7c7eb3654949\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.757187 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume\") pod \"7fddd336-9fea-4537-8aec-7c7eb3654949\" (UID: \"7fddd336-9fea-4537-8aec-7c7eb3654949\") " Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.758070 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume" (OuterVolumeSpecName: "config-volume") pod "7fddd336-9fea-4537-8aec-7c7eb3654949" (UID: "7fddd336-9fea-4537-8aec-7c7eb3654949"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.761176 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr" (OuterVolumeSpecName: "kube-api-access-vp7nr") pod "7fddd336-9fea-4537-8aec-7c7eb3654949" (UID: "7fddd336-9fea-4537-8aec-7c7eb3654949"). InnerVolumeSpecName "kube-api-access-vp7nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.764109 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7fddd336-9fea-4537-8aec-7c7eb3654949" (UID: "7fddd336-9fea-4537-8aec-7c7eb3654949"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.859041 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fddd336-9fea-4537-8aec-7c7eb3654949-config-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.859118 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fddd336-9fea-4537-8aec-7c7eb3654949-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.859133 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp7nr\" (UniqueName: \"kubernetes.io/projected/7fddd336-9fea-4537-8aec-7c7eb3654949-kube-api-access-vp7nr\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.963272 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" event={"ID":"7fddd336-9fea-4537-8aec-7c7eb3654949","Type":"ContainerDied","Data":"c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e"} Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.963323 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3f775b4912aec33755189bff58b675837f41effa9efb00be8b4419fd0269e8e" Jan 20 09:15:07 crc kubenswrapper[4967]: I0120 09:15:07.963338 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481675-bj9rd" Jan 20 09:15:08 crc kubenswrapper[4967]: I0120 09:15:08.974148 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"556459c1-0698-48cb-a63e-125c3f2b2641","Type":"ContainerStarted","Data":"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612"} Jan 20 09:15:08 crc kubenswrapper[4967]: I0120 09:15:08.991752 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keystone-kuttl-tests/openstackclient" podStartSLOduration=2.770544509 podStartE2EDuration="10.991733578s" podCreationTimestamp="2026-01-20 09:14:58 +0000 UTC" firstStartedPulling="2026-01-20 09:14:59.375931154 +0000 UTC m=+1533.961231361" lastFinishedPulling="2026-01-20 09:15:07.597120223 +0000 UTC m=+1542.182420430" observedRunningTime="2026-01-20 09:15:08.985563568 +0000 UTC m=+1543.570863785" watchObservedRunningTime="2026-01-20 09:15:08.991733578 +0000 UTC m=+1543.577033795" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.454438 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:13 crc kubenswrapper[4967]: E0120 09:15:13.455418 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fddd336-9fea-4537-8aec-7c7eb3654949" containerName="collect-profiles" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.455441 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fddd336-9fea-4537-8aec-7c7eb3654949" containerName="collect-profiles" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.455712 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fddd336-9fea-4537-8aec-7c7eb3654949" containerName="collect-profiles" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.457113 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.471505 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.636901 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-296n6\" (UniqueName: \"kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.636978 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.637018 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.738240 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-296n6\" (UniqueName: \"kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.738318 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.738359 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.738842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.738965 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.762698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-296n6\" (UniqueName: \"kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6\") pod \"redhat-marketplace-mhjvk\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:13 crc kubenswrapper[4967]: I0120 09:15:13.775878 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:14 crc kubenswrapper[4967]: I0120 09:15:14.014134 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:15 crc kubenswrapper[4967]: I0120 09:15:15.024735 4967 generic.go:334] "Generic (PLEG): container finished" podID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerID="bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809" exitCode=0 Jan 20 09:15:15 crc kubenswrapper[4967]: I0120 09:15:15.024851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerDied","Data":"bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809"} Jan 20 09:15:15 crc kubenswrapper[4967]: I0120 09:15:15.025122 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerStarted","Data":"ae8ffcc83fa6269a86bbbd5fddbb5186f78f1891e03dd5c396d14a5e922ee4fe"} Jan 20 09:15:17 crc kubenswrapper[4967]: I0120 09:15:17.052298 4967 generic.go:334] "Generic (PLEG): container finished" podID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerID="7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9" exitCode=0 Jan 20 09:15:17 crc kubenswrapper[4967]: I0120 09:15:17.052669 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerDied","Data":"7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9"} Jan 20 09:15:18 crc kubenswrapper[4967]: I0120 09:15:18.061086 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerStarted","Data":"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6"} Jan 20 09:15:18 crc kubenswrapper[4967]: I0120 09:15:18.074955 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mhjvk" podStartSLOduration=2.595640501 podStartE2EDuration="5.074933331s" podCreationTimestamp="2026-01-20 09:15:13 +0000 UTC" firstStartedPulling="2026-01-20 09:15:15.02728948 +0000 UTC m=+1549.612589687" lastFinishedPulling="2026-01-20 09:15:17.5065823 +0000 UTC m=+1552.091882517" observedRunningTime="2026-01-20 09:15:18.073567084 +0000 UTC m=+1552.658867311" watchObservedRunningTime="2026-01-20 09:15:18.074933331 +0000 UTC m=+1552.660233538" Jan 20 09:15:18 crc kubenswrapper[4967]: I0120 09:15:18.474401 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:15:18 crc kubenswrapper[4967]: I0120 09:15:18.474465 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.579317 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.583060 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.588113 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.720024 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m825z\" (UniqueName: \"kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.720118 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.720409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.822365 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m825z\" (UniqueName: \"kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.822506 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.822589 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.823127 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.823325 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.843019 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m825z\" (UniqueName: \"kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z\") pod \"redhat-operators-z59xr\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:19 crc kubenswrapper[4967]: I0120 09:15:19.910382 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:20 crc kubenswrapper[4967]: I0120 09:15:20.383402 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:20 crc kubenswrapper[4967]: W0120 09:15:20.390668 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45ab5049_ee16_4e37_871d_f52112152bd8.slice/crio-2b299f5b904abfc425535bca8ab0df32613e49f00837119c485b3fc57846292a WatchSource:0}: Error finding container 2b299f5b904abfc425535bca8ab0df32613e49f00837119c485b3fc57846292a: Status 404 returned error can't find the container with id 2b299f5b904abfc425535bca8ab0df32613e49f00837119c485b3fc57846292a Jan 20 09:15:21 crc kubenswrapper[4967]: I0120 09:15:21.092235 4967 generic.go:334] "Generic (PLEG): container finished" podID="45ab5049-ee16-4e37-871d-f52112152bd8" containerID="7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e" exitCode=0 Jan 20 09:15:21 crc kubenswrapper[4967]: I0120 09:15:21.092302 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerDied","Data":"7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e"} Jan 20 09:15:21 crc kubenswrapper[4967]: I0120 09:15:21.092568 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerStarted","Data":"2b299f5b904abfc425535bca8ab0df32613e49f00837119c485b3fc57846292a"} Jan 20 09:15:23 crc kubenswrapper[4967]: I0120 09:15:23.116303 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerStarted","Data":"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27"} Jan 20 09:15:23 crc kubenswrapper[4967]: I0120 09:15:23.776826 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:23 crc kubenswrapper[4967]: I0120 09:15:23.776869 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:23 crc kubenswrapper[4967]: I0120 09:15:23.884575 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:24 crc kubenswrapper[4967]: I0120 09:15:24.178442 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:24 crc kubenswrapper[4967]: I0120 09:15:24.755297 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:25 crc kubenswrapper[4967]: I0120 09:15:25.142995 4967 generic.go:334] "Generic (PLEG): container finished" podID="45ab5049-ee16-4e37-871d-f52112152bd8" containerID="652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27" exitCode=0 Jan 20 09:15:25 crc kubenswrapper[4967]: I0120 09:15:25.143689 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerDied","Data":"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27"} Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.186005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerStarted","Data":"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39"} Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.186203 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mhjvk" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="registry-server" containerID="cri-o://c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6" gracePeriod=2 Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.212817 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z59xr" podStartSLOduration=2.695726466 podStartE2EDuration="7.212789694s" podCreationTimestamp="2026-01-20 09:15:19 +0000 UTC" firstStartedPulling="2026-01-20 09:15:21.093600817 +0000 UTC m=+1555.678901024" lastFinishedPulling="2026-01-20 09:15:25.610664045 +0000 UTC m=+1560.195964252" observedRunningTime="2026-01-20 09:15:26.208123836 +0000 UTC m=+1560.793424043" watchObservedRunningTime="2026-01-20 09:15:26.212789694 +0000 UTC m=+1560.798089901" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.371948 4967 scope.go:117] "RemoveContainer" containerID="41b97d5eba37826f4dbc0832c212b967ba13596c23b3a4d9d292f447ca92f549" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.404568 4967 scope.go:117] "RemoveContainer" containerID="9a8d317eeb3fad645fc67e686323a9316759fa2785ce437b9e90e9f5f30854a9" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.634514 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.730081 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content\") pod \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.730147 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities\") pod \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.730264 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-296n6\" (UniqueName: \"kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6\") pod \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\" (UID: \"bbaa27c0-2a48-42f9-b242-f08dffc490ea\") " Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.732167 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities" (OuterVolumeSpecName: "utilities") pod "bbaa27c0-2a48-42f9-b242-f08dffc490ea" (UID: "bbaa27c0-2a48-42f9-b242-f08dffc490ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.736867 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6" (OuterVolumeSpecName: "kube-api-access-296n6") pod "bbaa27c0-2a48-42f9-b242-f08dffc490ea" (UID: "bbaa27c0-2a48-42f9-b242-f08dffc490ea"). InnerVolumeSpecName "kube-api-access-296n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.752292 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbaa27c0-2a48-42f9-b242-f08dffc490ea" (UID: "bbaa27c0-2a48-42f9-b242-f08dffc490ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.832543 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.832582 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbaa27c0-2a48-42f9-b242-f08dffc490ea-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:26 crc kubenswrapper[4967]: I0120 09:15:26.832595 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-296n6\" (UniqueName: \"kubernetes.io/projected/bbaa27c0-2a48-42f9-b242-f08dffc490ea-kube-api-access-296n6\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.194875 4967 generic.go:334] "Generic (PLEG): container finished" podID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerID="c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6" exitCode=0 Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.194919 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerDied","Data":"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6"} Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.194947 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mhjvk" event={"ID":"bbaa27c0-2a48-42f9-b242-f08dffc490ea","Type":"ContainerDied","Data":"ae8ffcc83fa6269a86bbbd5fddbb5186f78f1891e03dd5c396d14a5e922ee4fe"} Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.194968 4967 scope.go:117] "RemoveContainer" containerID="c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.194923 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mhjvk" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.227053 4967 scope.go:117] "RemoveContainer" containerID="7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.230238 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.236825 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mhjvk"] Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.251864 4967 scope.go:117] "RemoveContainer" containerID="bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.283402 4967 scope.go:117] "RemoveContainer" containerID="c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6" Jan 20 09:15:27 crc kubenswrapper[4967]: E0120 09:15:27.284009 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6\": container with ID starting with c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6 not found: ID does not exist" containerID="c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.284044 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6"} err="failed to get container status \"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6\": rpc error: code = NotFound desc = could not find container \"c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6\": container with ID starting with c7f73d7823301698a80d772000253583b86e275ed7430e3a5532fe0a3d66eca6 not found: ID does not exist" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.284081 4967 scope.go:117] "RemoveContainer" containerID="7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9" Jan 20 09:15:27 crc kubenswrapper[4967]: E0120 09:15:27.284438 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9\": container with ID starting with 7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9 not found: ID does not exist" containerID="7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.284469 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9"} err="failed to get container status \"7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9\": rpc error: code = NotFound desc = could not find container \"7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9\": container with ID starting with 7cc665d8eae4608ea984a26ab39ed31584e83a10e95cecff9a5f138e227eb5e9 not found: ID does not exist" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.284488 4967 scope.go:117] "RemoveContainer" containerID="bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809" Jan 20 09:15:27 crc kubenswrapper[4967]: E0120 09:15:27.284921 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809\": container with ID starting with bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809 not found: ID does not exist" containerID="bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.284973 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809"} err="failed to get container status \"bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809\": rpc error: code = NotFound desc = could not find container \"bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809\": container with ID starting with bf456ee07f9954b22c0d0419a77b6ebc8e2e8089af5a01dca345576d2fb23809 not found: ID does not exist" Jan 20 09:15:27 crc kubenswrapper[4967]: I0120 09:15:27.706055 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" path="/var/lib/kubelet/pods/bbaa27c0-2a48-42f9-b242-f08dffc490ea/volumes" Jan 20 09:15:29 crc kubenswrapper[4967]: I0120 09:15:29.910874 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:29 crc kubenswrapper[4967]: I0120 09:15:29.910959 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:31 crc kubenswrapper[4967]: I0120 09:15:31.085356 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z59xr" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="registry-server" probeResult="failure" output=< Jan 20 09:15:31 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 09:15:31 crc kubenswrapper[4967]: > Jan 20 09:15:39 crc kubenswrapper[4967]: I0120 09:15:39.949929 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:40 crc kubenswrapper[4967]: I0120 09:15:40.002096 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:40 crc kubenswrapper[4967]: I0120 09:15:40.183729 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.294687 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z59xr" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="registry-server" containerID="cri-o://a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39" gracePeriod=2 Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.711409 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.737267 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities\") pod \"45ab5049-ee16-4e37-871d-f52112152bd8\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.737325 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content\") pod \"45ab5049-ee16-4e37-871d-f52112152bd8\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.737473 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m825z\" (UniqueName: \"kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z\") pod \"45ab5049-ee16-4e37-871d-f52112152bd8\" (UID: \"45ab5049-ee16-4e37-871d-f52112152bd8\") " Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.738658 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities" (OuterVolumeSpecName: "utilities") pod "45ab5049-ee16-4e37-871d-f52112152bd8" (UID: "45ab5049-ee16-4e37-871d-f52112152bd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.747840 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z" (OuterVolumeSpecName: "kube-api-access-m825z") pod "45ab5049-ee16-4e37-871d-f52112152bd8" (UID: "45ab5049-ee16-4e37-871d-f52112152bd8"). InnerVolumeSpecName "kube-api-access-m825z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.839235 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m825z\" (UniqueName: \"kubernetes.io/projected/45ab5049-ee16-4e37-871d-f52112152bd8-kube-api-access-m825z\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.839285 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.872098 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45ab5049-ee16-4e37-871d-f52112152bd8" (UID: "45ab5049-ee16-4e37-871d-f52112152bd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:15:41 crc kubenswrapper[4967]: I0120 09:15:41.940573 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ab5049-ee16-4e37-871d-f52112152bd8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.303247 4967 generic.go:334] "Generic (PLEG): container finished" podID="45ab5049-ee16-4e37-871d-f52112152bd8" containerID="a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39" exitCode=0 Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.303287 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerDied","Data":"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39"} Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.303312 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z59xr" event={"ID":"45ab5049-ee16-4e37-871d-f52112152bd8","Type":"ContainerDied","Data":"2b299f5b904abfc425535bca8ab0df32613e49f00837119c485b3fc57846292a"} Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.303331 4967 scope.go:117] "RemoveContainer" containerID="a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.303337 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z59xr" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.319138 4967 scope.go:117] "RemoveContainer" containerID="652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.348679 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.351836 4967 scope.go:117] "RemoveContainer" containerID="7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.355972 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z59xr"] Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.380441 4967 scope.go:117] "RemoveContainer" containerID="a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39" Jan 20 09:15:42 crc kubenswrapper[4967]: E0120 09:15:42.381002 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39\": container with ID starting with a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39 not found: ID does not exist" containerID="a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.381053 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39"} err="failed to get container status \"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39\": rpc error: code = NotFound desc = could not find container \"a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39\": container with ID starting with a7b0648643eb56cc11a8a0bc992a287b42420af3b81126ad110579bc0a251f39 not found: ID does not exist" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.381084 4967 scope.go:117] "RemoveContainer" containerID="652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27" Jan 20 09:15:42 crc kubenswrapper[4967]: E0120 09:15:42.381572 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27\": container with ID starting with 652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27 not found: ID does not exist" containerID="652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.381677 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27"} err="failed to get container status \"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27\": rpc error: code = NotFound desc = could not find container \"652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27\": container with ID starting with 652889fac00f903ae980e287154b4d533f0913d0188f351523838c00982e1c27 not found: ID does not exist" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.381712 4967 scope.go:117] "RemoveContainer" containerID="7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e" Jan 20 09:15:42 crc kubenswrapper[4967]: E0120 09:15:42.382025 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e\": container with ID starting with 7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e not found: ID does not exist" containerID="7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e" Jan 20 09:15:42 crc kubenswrapper[4967]: I0120 09:15:42.382047 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e"} err="failed to get container status \"7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e\": rpc error: code = NotFound desc = could not find container \"7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e\": container with ID starting with 7aa61cad5732f37771953f2b8d91cdd2a4b95aa04fb3c4ce22cd7d8d38c4255e not found: ID does not exist" Jan 20 09:15:43 crc kubenswrapper[4967]: I0120 09:15:43.702068 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" path="/var/lib/kubelet/pods/45ab5049-ee16-4e37-871d-f52112152bd8/volumes" Jan 20 09:15:48 crc kubenswrapper[4967]: I0120 09:15:48.474375 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:15:48 crc kubenswrapper[4967]: I0120 09:15:48.474841 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:15:48 crc kubenswrapper[4967]: I0120 09:15:48.474905 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:15:48 crc kubenswrapper[4967]: I0120 09:15:48.476002 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:15:48 crc kubenswrapper[4967]: I0120 09:15:48.476241 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" gracePeriod=600 Jan 20 09:15:48 crc kubenswrapper[4967]: E0120 09:15:48.607816 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:15:49 crc kubenswrapper[4967]: I0120 09:15:49.361992 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" exitCode=0 Jan 20 09:15:49 crc kubenswrapper[4967]: I0120 09:15:49.362064 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2"} Jan 20 09:15:49 crc kubenswrapper[4967]: I0120 09:15:49.362282 4967 scope.go:117] "RemoveContainer" containerID="b9686d71d4a5ae3efdb5b3fff1d6f00de13b00dbccd72b46f4f7a73e9c72da1e" Jan 20 09:15:49 crc kubenswrapper[4967]: I0120 09:15:49.362973 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:15:49 crc kubenswrapper[4967]: E0120 09:15:49.363417 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:16:02 crc kubenswrapper[4967]: I0120 09:16:02.695249 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:16:02 crc kubenswrapper[4967]: E0120 09:16:02.696439 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:16:15 crc kubenswrapper[4967]: I0120 09:16:15.698816 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:16:15 crc kubenswrapper[4967]: E0120 09:16:15.699830 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:16:26 crc kubenswrapper[4967]: I0120 09:16:26.505187 4967 scope.go:117] "RemoveContainer" containerID="bc81d8e4fb4fae0f2bec4e802abc804c4b1f66342f49d35ae19abcc2704f9783" Jan 20 09:16:26 crc kubenswrapper[4967]: I0120 09:16:26.523177 4967 scope.go:117] "RemoveContainer" containerID="d0f8575ff69255bd79a19bdb792557eebfc4e793602739a31da09252c42c0062" Jan 20 09:16:26 crc kubenswrapper[4967]: I0120 09:16:26.562995 4967 scope.go:117] "RemoveContainer" containerID="911d966924f4b1c670c0c0ca4760ce4e647853a40d3f2c6f0e4f12830e38dcdf" Jan 20 09:16:26 crc kubenswrapper[4967]: I0120 09:16:26.694433 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:16:26 crc kubenswrapper[4967]: E0120 09:16:26.694693 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:16:39 crc kubenswrapper[4967]: I0120 09:16:39.695153 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:16:39 crc kubenswrapper[4967]: E0120 09:16:39.695945 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:16:51 crc kubenswrapper[4967]: I0120 09:16:51.714944 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:16:51 crc kubenswrapper[4967]: E0120 09:16:51.715945 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:17:03 crc kubenswrapper[4967]: I0120 09:17:03.698917 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:17:03 crc kubenswrapper[4967]: E0120 09:17:03.699814 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:17:18 crc kubenswrapper[4967]: I0120 09:17:18.694516 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:17:18 crc kubenswrapper[4967]: E0120 09:17:18.695715 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:17:26 crc kubenswrapper[4967]: I0120 09:17:26.659134 4967 scope.go:117] "RemoveContainer" containerID="c03e74ad0f638055038a7d4eacc9865025f4db5e78781645cb989600330b1b6f" Jan 20 09:17:26 crc kubenswrapper[4967]: I0120 09:17:26.678434 4967 scope.go:117] "RemoveContainer" containerID="2012199ae0c330813c9c0e10ae32e0a6ed11e9b92e0f92b36312b89f1cf45850" Jan 20 09:17:26 crc kubenswrapper[4967]: I0120 09:17:26.718338 4967 scope.go:117] "RemoveContainer" containerID="ce2532cd2d4d6d323ced22f0fbb751a8b077e9a85baf051ed58d836ed424b1f2" Jan 20 09:17:26 crc kubenswrapper[4967]: I0120 09:17:26.748183 4967 scope.go:117] "RemoveContainer" containerID="dd2084d9acd3ff737c39b495b30363e0731669a0f84d5dfee73a6fc3aa26e5d5" Jan 20 09:17:29 crc kubenswrapper[4967]: I0120 09:17:29.694551 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:17:29 crc kubenswrapper[4967]: E0120 09:17:29.695064 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:17:40 crc kubenswrapper[4967]: I0120 09:17:40.693775 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:17:40 crc kubenswrapper[4967]: E0120 09:17:40.695354 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:17:53 crc kubenswrapper[4967]: I0120 09:17:53.694836 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:17:53 crc kubenswrapper[4967]: E0120 09:17:53.697979 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:18:07 crc kubenswrapper[4967]: I0120 09:18:07.694732 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:18:07 crc kubenswrapper[4967]: E0120 09:18:07.695599 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:18:14 crc kubenswrapper[4967]: I0120 09:18:14.057675 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-jffl5"] Jan 20 09:18:14 crc kubenswrapper[4967]: I0120 09:18:14.061641 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-jffl5"] Jan 20 09:18:15 crc kubenswrapper[4967]: I0120 09:18:15.704214 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="135dd5ea-bbbd-4412-a188-8725e5dc8ee5" path="/var/lib/kubelet/pods/135dd5ea-bbbd-4412-a188-8725e5dc8ee5/volumes" Jan 20 09:18:21 crc kubenswrapper[4967]: I0120 09:18:21.694584 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:18:21 crc kubenswrapper[4967]: E0120 09:18:21.695093 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.809267 4967 scope.go:117] "RemoveContainer" containerID="0bd1d7fe79bc8def86b98d6b9d9262b7287d0c38474b0e39e3784e72af8f5504" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.853914 4967 scope.go:117] "RemoveContainer" containerID="3f4434d588871cdbeb73529bf7aea288ba1bad8370ad11029c6fe8bb7580ad0d" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.884947 4967 scope.go:117] "RemoveContainer" containerID="c878c6bd96ecf98696e86ce007b6150e1ae1801a931f59eba063a3d3efbc5d9b" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.913437 4967 scope.go:117] "RemoveContainer" containerID="2af072d4195b894d117a42c260d7e2909a983afe4638c2ccfa8358b2731e581d" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.934259 4967 scope.go:117] "RemoveContainer" containerID="eff3a2e2f9d176b6a11fb1e700aa1d2d3dfe26ef48446d0193fef0f1c8ced2c0" Jan 20 09:18:26 crc kubenswrapper[4967]: I0120 09:18:26.958620 4967 scope.go:117] "RemoveContainer" containerID="5e764437a596ed3a0227ac146257586716ffabeb38e8253b06740f511b6b411a" Jan 20 09:18:33 crc kubenswrapper[4967]: I0120 09:18:33.696472 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:18:33 crc kubenswrapper[4967]: E0120 09:18:33.698353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:18:47 crc kubenswrapper[4967]: I0120 09:18:47.695451 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:18:47 crc kubenswrapper[4967]: E0120 09:18:47.697943 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:19:00 crc kubenswrapper[4967]: I0120 09:19:00.695109 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:19:00 crc kubenswrapper[4967]: E0120 09:19:00.696343 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:19:13 crc kubenswrapper[4967]: I0120 09:19:13.694333 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:19:13 crc kubenswrapper[4967]: E0120 09:19:13.695149 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:19:26 crc kubenswrapper[4967]: I0120 09:19:26.694067 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:19:26 crc kubenswrapper[4967]: E0120 09:19:26.694783 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:19:27 crc kubenswrapper[4967]: I0120 09:19:27.037203 4967 scope.go:117] "RemoveContainer" containerID="47890fa2d12fc82ecec684c9ada7a1689722c0b13db888e00d5757865b74a923" Jan 20 09:19:27 crc kubenswrapper[4967]: I0120 09:19:27.075288 4967 scope.go:117] "RemoveContainer" containerID="f1256ed292eae1a7d754635ae38ca6e2eb72c33afa8fde8ef55b5ead4073ee5e" Jan 20 09:19:27 crc kubenswrapper[4967]: I0120 09:19:27.127814 4967 scope.go:117] "RemoveContainer" containerID="2e1fefd0f41067fcb00e4843ac0f79abf3545a40d39bb1f9a31a6d11939fa584" Jan 20 09:19:27 crc kubenswrapper[4967]: I0120 09:19:27.148378 4967 scope.go:117] "RemoveContainer" containerID="85c48fe3b63a501efb76238b7815533a6da2b2df72897ff06fb2db5198706593" Jan 20 09:19:27 crc kubenswrapper[4967]: I0120 09:19:27.173634 4967 scope.go:117] "RemoveContainer" containerID="d475f9b8498897f9c18a47b5ac0b581c2fa8d09b9c1db4f538c2d70b2795c357" Jan 20 09:19:40 crc kubenswrapper[4967]: I0120 09:19:40.693756 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:19:40 crc kubenswrapper[4967]: E0120 09:19:40.694375 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:19:52 crc kubenswrapper[4967]: I0120 09:19:52.694483 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:19:52 crc kubenswrapper[4967]: E0120 09:19:52.695279 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:20:07 crc kubenswrapper[4967]: I0120 09:20:07.694827 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:20:07 crc kubenswrapper[4967]: E0120 09:20:07.697145 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:20:18 crc kubenswrapper[4967]: I0120 09:20:18.694338 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:20:18 crc kubenswrapper[4967]: E0120 09:20:18.694874 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:20:27 crc kubenswrapper[4967]: I0120 09:20:27.295794 4967 scope.go:117] "RemoveContainer" containerID="e9efeeafcd07e757fe67dc9b21d531b2b08b8097d021a9dd6b63c1ee0dea12f2" Jan 20 09:20:32 crc kubenswrapper[4967]: I0120 09:20:32.693704 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:20:32 crc kubenswrapper[4967]: E0120 09:20:32.694458 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:20:46 crc kubenswrapper[4967]: I0120 09:20:46.694631 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:20:46 crc kubenswrapper[4967]: E0120 09:20:46.695375 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:21:00 crc kubenswrapper[4967]: I0120 09:21:00.693643 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:21:01 crc kubenswrapper[4967]: I0120 09:21:01.795007 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc"} Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.472972 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473727 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="extract-utilities" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473742 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="extract-utilities" Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473757 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="extract-content" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473767 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="extract-content" Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473784 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="extract-utilities" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473793 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="extract-utilities" Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473805 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473813 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473833 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="extract-content" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473842 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="extract-content" Jan 20 09:21:22 crc kubenswrapper[4967]: E0120 09:21:22.473856 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.473864 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.474007 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="45ab5049-ee16-4e37-871d-f52112152bd8" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.474028 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbaa27c0-2a48-42f9-b242-f08dffc490ea" containerName="registry-server" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.475223 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.492357 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.623848 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.623891 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.624053 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffgq5\" (UniqueName: \"kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.725259 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.725334 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.725533 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffgq5\" (UniqueName: \"kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.726124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.726153 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.748914 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffgq5\" (UniqueName: \"kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5\") pod \"certified-operators-mxnb9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:22 crc kubenswrapper[4967]: I0120 09:21:22.815164 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:23 crc kubenswrapper[4967]: I0120 09:21:23.074073 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:23 crc kubenswrapper[4967]: I0120 09:21:23.947539 4967 generic.go:334] "Generic (PLEG): container finished" podID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerID="38f3acc56778cb043ed80a8371ef3b6f9ea21c07c29755159d817192250be734" exitCode=0 Jan 20 09:21:23 crc kubenswrapper[4967]: I0120 09:21:23.947684 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerDied","Data":"38f3acc56778cb043ed80a8371ef3b6f9ea21c07c29755159d817192250be734"} Jan 20 09:21:23 crc kubenswrapper[4967]: I0120 09:21:23.948005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerStarted","Data":"9222622fd809af459bc997341c1b37a1805e2443867baf6fb6d9549bbe5924d5"} Jan 20 09:21:23 crc kubenswrapper[4967]: I0120 09:21:23.950529 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 09:21:25 crc kubenswrapper[4967]: I0120 09:21:25.962372 4967 generic.go:334] "Generic (PLEG): container finished" podID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerID="5dcc481c31547148f0dd6b1862fc6cf7a65c55316f0d5afbae0c1fa89032d58d" exitCode=0 Jan 20 09:21:25 crc kubenswrapper[4967]: I0120 09:21:25.962428 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerDied","Data":"5dcc481c31547148f0dd6b1862fc6cf7a65c55316f0d5afbae0c1fa89032d58d"} Jan 20 09:21:26 crc kubenswrapper[4967]: I0120 09:21:26.973554 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerStarted","Data":"c8078d5bbb94994124d12630683c46ed1c57acd01bce9e30e60f9e2a4b2f9833"} Jan 20 09:21:26 crc kubenswrapper[4967]: I0120 09:21:26.997314 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mxnb9" podStartSLOduration=2.220094825 podStartE2EDuration="4.997285827s" podCreationTimestamp="2026-01-20 09:21:22 +0000 UTC" firstStartedPulling="2026-01-20 09:21:23.949945761 +0000 UTC m=+1918.535246008" lastFinishedPulling="2026-01-20 09:21:26.727136803 +0000 UTC m=+1921.312437010" observedRunningTime="2026-01-20 09:21:26.993945556 +0000 UTC m=+1921.579245833" watchObservedRunningTime="2026-01-20 09:21:26.997285827 +0000 UTC m=+1921.582586104" Jan 20 09:21:32 crc kubenswrapper[4967]: I0120 09:21:32.816121 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:32 crc kubenswrapper[4967]: I0120 09:21:32.817814 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:32 crc kubenswrapper[4967]: I0120 09:21:32.866091 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:33 crc kubenswrapper[4967]: I0120 09:21:33.059875 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:33 crc kubenswrapper[4967]: I0120 09:21:33.109942 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:35 crc kubenswrapper[4967]: I0120 09:21:35.028775 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mxnb9" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="registry-server" containerID="cri-o://c8078d5bbb94994124d12630683c46ed1c57acd01bce9e30e60f9e2a4b2f9833" gracePeriod=2 Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.037578 4967 generic.go:334] "Generic (PLEG): container finished" podID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerID="c8078d5bbb94994124d12630683c46ed1c57acd01bce9e30e60f9e2a4b2f9833" exitCode=0 Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.037765 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerDied","Data":"c8078d5bbb94994124d12630683c46ed1c57acd01bce9e30e60f9e2a4b2f9833"} Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.093761 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.255931 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content\") pod \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.255998 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities\") pod \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.256145 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffgq5\" (UniqueName: \"kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5\") pod \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\" (UID: \"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9\") " Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.256891 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities" (OuterVolumeSpecName: "utilities") pod "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" (UID: "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.260897 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5" (OuterVolumeSpecName: "kube-api-access-ffgq5") pod "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" (UID: "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9"). InnerVolumeSpecName "kube-api-access-ffgq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.298494 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" (UID: "ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.357897 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffgq5\" (UniqueName: \"kubernetes.io/projected/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-kube-api-access-ffgq5\") on node \"crc\" DevicePath \"\"" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.357960 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:21:36 crc kubenswrapper[4967]: I0120 09:21:36.357973 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.048743 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxnb9" event={"ID":"ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9","Type":"ContainerDied","Data":"9222622fd809af459bc997341c1b37a1805e2443867baf6fb6d9549bbe5924d5"} Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.048816 4967 scope.go:117] "RemoveContainer" containerID="c8078d5bbb94994124d12630683c46ed1c57acd01bce9e30e60f9e2a4b2f9833" Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.048878 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxnb9" Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.072624 4967 scope.go:117] "RemoveContainer" containerID="5dcc481c31547148f0dd6b1862fc6cf7a65c55316f0d5afbae0c1fa89032d58d" Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.099729 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.106812 4967 scope.go:117] "RemoveContainer" containerID="38f3acc56778cb043ed80a8371ef3b6f9ea21c07c29755159d817192250be734" Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.111470 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mxnb9"] Jan 20 09:21:37 crc kubenswrapper[4967]: I0120 09:21:37.703768 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" path="/var/lib/kubelet/pods/ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9/volumes" Jan 20 09:23:18 crc kubenswrapper[4967]: I0120 09:23:18.474536 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:23:18 crc kubenswrapper[4967]: I0120 09:23:18.475135 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:23:48 crc kubenswrapper[4967]: I0120 09:23:48.474957 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:23:48 crc kubenswrapper[4967]: I0120 09:23:48.475695 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.049327 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-qv4qt"] Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.059115 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-c611-account-create-update-glt4n"] Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.067882 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-create-qv4qt"] Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.075192 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-c611-account-create-update-glt4n"] Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.708241 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb7901f-4270-4905-a23b-1dba2d205315" path="/var/lib/kubelet/pods/1bb7901f-4270-4905-a23b-1dba2d205315/volumes" Jan 20 09:24:11 crc kubenswrapper[4967]: I0120 09:24:11.709368 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99758e56-9a70-42da-af60-10cd9ffcacad" path="/var/lib/kubelet/pods/99758e56-9a70-42da-af60-10cd9ffcacad/volumes" Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.033869 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f6bfl"] Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.040008 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-db-sync-f6bfl"] Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.474953 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.475019 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.475083 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.475845 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.475932 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc" gracePeriod=600 Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.604982 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc" exitCode=0 Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.605122 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc"} Jan 20 09:24:18 crc kubenswrapper[4967]: I0120 09:24:18.605338 4967 scope.go:117] "RemoveContainer" containerID="c2d3dbd5a8c50f93696daf2877708cfedab872729fff815e52781a8aadc345b2" Jan 20 09:24:19 crc kubenswrapper[4967]: I0120 09:24:19.618065 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723"} Jan 20 09:24:19 crc kubenswrapper[4967]: I0120 09:24:19.705026 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3c6996-85b2-42d6-879b-f4c3acbe6e75" path="/var/lib/kubelet/pods/ee3c6996-85b2-42d6-879b-f4c3acbe6e75/volumes" Jan 20 09:24:25 crc kubenswrapper[4967]: I0120 09:24:25.047886 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-755cf"] Jan 20 09:24:25 crc kubenswrapper[4967]: I0120 09:24:25.063991 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-bootstrap-755cf"] Jan 20 09:24:25 crc kubenswrapper[4967]: I0120 09:24:25.704025 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d43314-9260-4e86-9003-6bd022c9c390" path="/var/lib/kubelet/pods/c8d43314-9260-4e86-9003-6bd022c9c390/volumes" Jan 20 09:24:27 crc kubenswrapper[4967]: I0120 09:24:27.418855 4967 scope.go:117] "RemoveContainer" containerID="dede0cf90ddacbad57128ccc6624702d6f2d87714b4d5d531248a9f2c7459ffa" Jan 20 09:24:27 crc kubenswrapper[4967]: I0120 09:24:27.457276 4967 scope.go:117] "RemoveContainer" containerID="4ea689945fbcce99f3caa54beeceab17153ccef9a71af7cce581407b4be2b0c4" Jan 20 09:24:27 crc kubenswrapper[4967]: I0120 09:24:27.478403 4967 scope.go:117] "RemoveContainer" containerID="e4060a98764541a9fe8ea245229b415f396c644ccb527a93f49ec8debf047923" Jan 20 09:24:27 crc kubenswrapper[4967]: I0120 09:24:27.527179 4967 scope.go:117] "RemoveContainer" containerID="c9ebcb9214c544fcaf6c7da0c9092d0365c183dcbd8d19a109313de62197d4b0" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.433632 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.434464 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstackclient" podUID="556459c1-0698-48cb-a63e-125c3f2b2641" containerName="openstackclient" containerID="cri-o://ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612" gracePeriod=30 Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.856696 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.954479 4967 generic.go:334] "Generic (PLEG): container finished" podID="556459c1-0698-48cb-a63e-125c3f2b2641" containerID="ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612" exitCode=143 Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.954522 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"556459c1-0698-48cb-a63e-125c3f2b2641","Type":"ContainerDied","Data":"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612"} Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.954547 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstackclient" event={"ID":"556459c1-0698-48cb-a63e-125c3f2b2641","Type":"ContainerDied","Data":"300511ee8d0b5f0c14b57058580fa7bfc99df23200afbfc27b62822f38117869"} Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.954563 4967 scope.go:117] "RemoveContainer" containerID="ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.954567 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstackclient" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.972940 4967 scope.go:117] "RemoveContainer" containerID="ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612" Jan 20 09:25:14 crc kubenswrapper[4967]: E0120 09:25:14.973541 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612\": container with ID starting with ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612 not found: ID does not exist" containerID="ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.973580 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612"} err="failed to get container status \"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612\": rpc error: code = NotFound desc = could not find container \"ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612\": container with ID starting with ccde50b8ef02966e3abe78a9038eb963518d1839aa22c34f6e874ef5fa298612 not found: ID does not exist" Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.994419 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config\") pod \"556459c1-0698-48cb-a63e-125c3f2b2641\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.994791 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5mx8\" (UniqueName: \"kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8\") pod \"556459c1-0698-48cb-a63e-125c3f2b2641\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " Jan 20 09:25:14 crc kubenswrapper[4967]: I0120 09:25:14.994821 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret\") pod \"556459c1-0698-48cb-a63e-125c3f2b2641\" (UID: \"556459c1-0698-48cb-a63e-125c3f2b2641\") " Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.003312 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8" (OuterVolumeSpecName: "kube-api-access-s5mx8") pod "556459c1-0698-48cb-a63e-125c3f2b2641" (UID: "556459c1-0698-48cb-a63e-125c3f2b2641"). InnerVolumeSpecName "kube-api-access-s5mx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.010996 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "556459c1-0698-48cb-a63e-125c3f2b2641" (UID: "556459c1-0698-48cb-a63e-125c3f2b2641"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.012860 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "556459c1-0698-48cb-a63e-125c3f2b2641" (UID: "556459c1-0698-48cb-a63e-125c3f2b2641"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.096949 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5mx8\" (UniqueName: \"kubernetes.io/projected/556459c1-0698-48cb-a63e-125c3f2b2641-kube-api-access-s5mx8\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.096998 4967 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.097012 4967 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/556459c1-0698-48cb-a63e-125c3f2b2641-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.288420 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.292840 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstackclient"] Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.431554 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.432039 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" podUID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" containerName="keystone-api" containerID="cri-o://278537679b0b42a5c14d897c47331896aed384ed89feda18d3ccd00bd94f9416" gracePeriod=30 Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.502593 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/keystonec611-account-delete-gssrv"] Jan 20 09:25:15 crc kubenswrapper[4967]: E0120 09:25:15.503936 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556459c1-0698-48cb-a63e-125c3f2b2641" containerName="openstackclient" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504037 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="556459c1-0698-48cb-a63e-125c3f2b2641" containerName="openstackclient" Jan 20 09:25:15 crc kubenswrapper[4967]: E0120 09:25:15.504102 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="extract-content" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504165 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="extract-content" Jan 20 09:25:15 crc kubenswrapper[4967]: E0120 09:25:15.504235 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="registry-server" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504299 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="registry-server" Jan 20 09:25:15 crc kubenswrapper[4967]: E0120 09:25:15.504357 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="extract-utilities" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504420 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="extract-utilities" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504605 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="556459c1-0698-48cb-a63e-125c3f2b2641" containerName="openstackclient" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.504697 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebbb40fd-e4a5-40d8-9cd7-fdfc3a1174d9" containerName="registry-server" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.505334 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.509905 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonec611-account-delete-gssrv"] Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.603861 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dcvj\" (UniqueName: \"kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.604045 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.701983 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="556459c1-0698-48cb-a63e-125c3f2b2641" path="/var/lib/kubelet/pods/556459c1-0698-48cb-a63e-125c3f2b2641/volumes" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.705755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.705814 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dcvj\" (UniqueName: \"kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.706862 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.725946 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dcvj\" (UniqueName: \"kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj\") pod \"keystonec611-account-delete-gssrv\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:15 crc kubenswrapper[4967]: I0120 09:25:15.822918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:16 crc kubenswrapper[4967]: I0120 09:25:16.234736 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/keystonec611-account-delete-gssrv"] Jan 20 09:25:16 crc kubenswrapper[4967]: I0120 09:25:16.972158 4967 generic.go:334] "Generic (PLEG): container finished" podID="cdc1a322-823c-4011-8b16-d62e43342149" containerID="c8333d326331a494cdd703a528b99b8a1c57448d6654d981371357ccda60a8e6" exitCode=0 Jan 20 09:25:16 crc kubenswrapper[4967]: I0120 09:25:16.972252 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" event={"ID":"cdc1a322-823c-4011-8b16-d62e43342149","Type":"ContainerDied","Data":"c8333d326331a494cdd703a528b99b8a1c57448d6654d981371357ccda60a8e6"} Jan 20 09:25:16 crc kubenswrapper[4967]: I0120 09:25:16.972435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" event={"ID":"cdc1a322-823c-4011-8b16-d62e43342149","Type":"ContainerStarted","Data":"65fa93cfd6f20cd1697b3b5b6826c09c356fc3d5531253287f85f6990820e45c"} Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.237227 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.335627 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts\") pod \"cdc1a322-823c-4011-8b16-d62e43342149\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.335834 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dcvj\" (UniqueName: \"kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj\") pod \"cdc1a322-823c-4011-8b16-d62e43342149\" (UID: \"cdc1a322-823c-4011-8b16-d62e43342149\") " Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.336605 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cdc1a322-823c-4011-8b16-d62e43342149" (UID: "cdc1a322-823c-4011-8b16-d62e43342149"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.340306 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj" (OuterVolumeSpecName: "kube-api-access-5dcvj") pod "cdc1a322-823c-4011-8b16-d62e43342149" (UID: "cdc1a322-823c-4011-8b16-d62e43342149"). InnerVolumeSpecName "kube-api-access-5dcvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.437796 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dcvj\" (UniqueName: \"kubernetes.io/projected/cdc1a322-823c-4011-8b16-d62e43342149-kube-api-access-5dcvj\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.437833 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdc1a322-823c-4011-8b16-d62e43342149-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.989242 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.989242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystonec611-account-delete-gssrv" event={"ID":"cdc1a322-823c-4011-8b16-d62e43342149","Type":"ContainerDied","Data":"65fa93cfd6f20cd1697b3b5b6826c09c356fc3d5531253287f85f6990820e45c"} Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.989808 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65fa93cfd6f20cd1697b3b5b6826c09c356fc3d5531253287f85f6990820e45c" Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.990887 4967 generic.go:334] "Generic (PLEG): container finished" podID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" containerID="278537679b0b42a5c14d897c47331896aed384ed89feda18d3ccd00bd94f9416" exitCode=0 Jan 20 09:25:18 crc kubenswrapper[4967]: I0120 09:25:18.990925 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" event={"ID":"8e0e5e9b-0600-403a-a4b6-ecb254da3396","Type":"ContainerDied","Data":"278537679b0b42a5c14d897c47331896aed384ed89feda18d3ccd00bd94f9416"} Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.354427 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.451359 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys\") pod \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.451412 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts\") pod \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.451441 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys\") pod \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.451512 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pvzh\" (UniqueName: \"kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh\") pod \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.451568 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data\") pod \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\" (UID: \"8e0e5e9b-0600-403a-a4b6-ecb254da3396\") " Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.455827 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh" (OuterVolumeSpecName: "kube-api-access-9pvzh") pod "8e0e5e9b-0600-403a-a4b6-ecb254da3396" (UID: "8e0e5e9b-0600-403a-a4b6-ecb254da3396"). InnerVolumeSpecName "kube-api-access-9pvzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.456533 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8e0e5e9b-0600-403a-a4b6-ecb254da3396" (UID: "8e0e5e9b-0600-403a-a4b6-ecb254da3396"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.456913 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts" (OuterVolumeSpecName: "scripts") pod "8e0e5e9b-0600-403a-a4b6-ecb254da3396" (UID: "8e0e5e9b-0600-403a-a4b6-ecb254da3396"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.456936 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8e0e5e9b-0600-403a-a4b6-ecb254da3396" (UID: "8e0e5e9b-0600-403a-a4b6-ecb254da3396"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.478831 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data" (OuterVolumeSpecName: "config-data") pod "8e0e5e9b-0600-403a-a4b6-ecb254da3396" (UID: "8e0e5e9b-0600-403a-a4b6-ecb254da3396"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.552583 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.552643 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.552653 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.552662 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pvzh\" (UniqueName: \"kubernetes.io/projected/8e0e5e9b-0600-403a-a4b6-ecb254da3396-kube-api-access-9pvzh\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.552674 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e0e5e9b-0600-403a-a4b6-ecb254da3396-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.997944 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" event={"ID":"8e0e5e9b-0600-403a-a4b6-ecb254da3396","Type":"ContainerDied","Data":"8efb0f9f0b899016b17879e45ec51300c3fcd4c2f95a390e34fe7e593ee76aed"} Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.997982 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/keystone-7d48b4856d-x6ktr" Jan 20 09:25:19 crc kubenswrapper[4967]: I0120 09:25:19.998280 4967 scope.go:117] "RemoveContainer" containerID="278537679b0b42a5c14d897c47331896aed384ed89feda18d3ccd00bd94f9416" Jan 20 09:25:20 crc kubenswrapper[4967]: I0120 09:25:20.014217 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:25:20 crc kubenswrapper[4967]: I0120 09:25:20.018296 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystone-7d48b4856d-x6ktr"] Jan 20 09:25:20 crc kubenswrapper[4967]: I0120 09:25:20.538421 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/keystonec611-account-delete-gssrv"] Jan 20 09:25:20 crc kubenswrapper[4967]: I0120 09:25:20.542774 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/keystonec611-account-delete-gssrv"] Jan 20 09:25:21 crc kubenswrapper[4967]: I0120 09:25:21.709310 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" path="/var/lib/kubelet/pods/8e0e5e9b-0600-403a-a4b6-ecb254da3396/volumes" Jan 20 09:25:21 crc kubenswrapper[4967]: I0120 09:25:21.710466 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc1a322-823c-4011-8b16-d62e43342149" path="/var/lib/kubelet/pods/cdc1a322-823c-4011-8b16-d62e43342149/volumes" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.782913 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["keystone-kuttl-tests/root-account-create-update-gj5qg"] Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.784230 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" containerName="keystone-api" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.784248 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" containerName="keystone-api" Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.784273 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc1a322-823c-4011-8b16-d62e43342149" containerName="mariadb-account-delete" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.784279 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc1a322-823c-4011-8b16-d62e43342149" containerName="mariadb-account-delete" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.784416 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0e5e9b-0600-403a-a4b6-ecb254da3396" containerName="keystone-api" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.784427 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc1a322-823c-4011-8b16-d62e43342149" containerName="mariadb-account-delete" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.785156 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.795030 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-gj5qg"] Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.797970 4967 reflector.go:368] Caches populated for *v1.Secret from object-"keystone-kuttl-tests"/"openstack-mariadb-root-db-secret" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.820816 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.834669 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.841637 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.857306 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-gj5qg"] Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.857970 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-8n2sq operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="keystone-kuttl-tests/root-account-create-update-gj5qg" podUID="a324d526-a683-439e-8d0a-24ad6dfb3b57" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.862420 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.862509 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n2sq\" (UniqueName: \"kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.963439 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.963509 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n2sq\" (UniqueName: \"kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.963622 4967 configmap.go:193] Couldn't get configMap keystone-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.963712 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts podName:a324d526-a683-439e-8d0a-24ad6dfb3b57 nodeName:}" failed. No retries permitted until 2026-01-20 09:25:28.463689847 +0000 UTC m=+2163.048990124 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts") pod "root-account-create-update-gj5qg" (UID: "a324d526-a683-439e-8d0a-24ad6dfb3b57") : configmap "openstack-scripts" not found Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.969021 4967 projected.go:194] Error preparing data for projected volume kube-api-access-8n2sq for pod keystone-kuttl-tests/root-account-create-update-gj5qg: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 20 09:25:27 crc kubenswrapper[4967]: E0120 09:25:27.969096 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq podName:a324d526-a683-439e-8d0a-24ad6dfb3b57 nodeName:}" failed. No retries permitted until 2026-01-20 09:25:28.469078094 +0000 UTC m=+2163.054378301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8n2sq" (UniqueName: "kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq") pod "root-account-create-update-gj5qg" (UID: "a324d526-a683-439e-8d0a-24ad6dfb3b57") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 20 09:25:27 crc kubenswrapper[4967]: I0120 09:25:27.982079 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-2" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="galera" containerID="cri-o://ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92" gracePeriod=30 Jan 20 09:25:28 crc kubenswrapper[4967]: I0120 09:25:28.053221 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:28 crc kubenswrapper[4967]: I0120 09:25:28.059766 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:28 crc kubenswrapper[4967]: I0120 09:25:28.470261 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n2sq\" (UniqueName: \"kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:28 crc kubenswrapper[4967]: I0120 09:25:28.470601 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts\") pod \"root-account-create-update-gj5qg\" (UID: \"a324d526-a683-439e-8d0a-24ad6dfb3b57\") " pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:28 crc kubenswrapper[4967]: E0120 09:25:28.470838 4967 configmap.go:193] Couldn't get configMap keystone-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 20 09:25:28 crc kubenswrapper[4967]: E0120 09:25:28.471028 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts podName:a324d526-a683-439e-8d0a-24ad6dfb3b57 nodeName:}" failed. No retries permitted until 2026-01-20 09:25:29.470986987 +0000 UTC m=+2164.056287214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts") pod "root-account-create-update-gj5qg" (UID: "a324d526-a683-439e-8d0a-24ad6dfb3b57") : configmap "openstack-scripts" not found Jan 20 09:25:28 crc kubenswrapper[4967]: E0120 09:25:28.473669 4967 projected.go:194] Error preparing data for projected volume kube-api-access-8n2sq for pod keystone-kuttl-tests/root-account-create-update-gj5qg: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 20 09:25:28 crc kubenswrapper[4967]: E0120 09:25:28.473722 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq podName:a324d526-a683-439e-8d0a-24ad6dfb3b57 nodeName:}" failed. No retries permitted until 2026-01-20 09:25:29.473711671 +0000 UTC m=+2164.059011878 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-8n2sq" (UniqueName: "kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq") pod "root-account-create-update-gj5qg" (UID: "a324d526-a683-439e-8d0a-24ad6dfb3b57") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.060226 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/root-account-create-update-gj5qg" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.129558 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-gj5qg"] Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.138239 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/root-account-create-update-gj5qg"] Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.163332 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.163577 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/memcached-0" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerName="memcached" containerID="cri-o://5a745e943b60dccc0dcba24f62de592b210fe6ab152e694350a3f95b818fccf9" gracePeriod=30 Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.281089 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n2sq\" (UniqueName: \"kubernetes.io/projected/a324d526-a683-439e-8d0a-24ad6dfb3b57-kube-api-access-8n2sq\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.281128 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a324d526-a683-439e-8d0a-24ad6dfb3b57-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.563974 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.643092 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691493 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691548 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691639 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691676 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65kbt\" (UniqueName: \"kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691713 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.691767 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated\") pod \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\" (UID: \"689be023-b3b2-4b2c-b16e-5a24ce36edfc\") " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.692855 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.692959 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.693382 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.693495 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.698748 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt" (OuterVolumeSpecName: "kube-api-access-65kbt") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "kube-api-access-65kbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.709604 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a324d526-a683-439e-8d0a-24ad6dfb3b57" path="/var/lib/kubelet/pods/a324d526-a683-439e-8d0a-24ad6dfb3b57/volumes" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.716772 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "689be023-b3b2-4b2c-b16e-5a24ce36edfc" (UID: "689be023-b3b2-4b2c-b16e-5a24ce36edfc"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793254 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793299 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793312 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65kbt\" (UniqueName: \"kubernetes.io/projected/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kube-api-access-65kbt\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793326 4967 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793336 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.793349 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689be023-b3b2-4b2c-b16e-5a24ce36edfc-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.803532 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.894813 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.926894 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:25:29 crc kubenswrapper[4967]: I0120 09:25:29.990764 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-1" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="galera" containerID="cri-o://75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" gracePeriod=28 Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.068029 4967 generic.go:334] "Generic (PLEG): container finished" podID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerID="ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92" exitCode=0 Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.068291 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-2" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.071934 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerDied","Data":"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92"} Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.072030 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-2" event={"ID":"689be023-b3b2-4b2c-b16e-5a24ce36edfc","Type":"ContainerDied","Data":"375d2ac1b5cd2c9bd022f3db354145e25ce423f089e43fa09e9cd0ca66e560c6"} Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.072052 4967 scope.go:117] "RemoveContainer" containerID="ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.094730 4967 scope.go:117] "RemoveContainer" containerID="2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.098733 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.107247 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-2"] Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.113079 4967 scope.go:117] "RemoveContainer" containerID="ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92" Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.113563 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92\": container with ID starting with ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92 not found: ID does not exist" containerID="ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.113719 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92"} err="failed to get container status \"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92\": rpc error: code = NotFound desc = could not find container \"ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92\": container with ID starting with ebcc3c3813eb476b3c068ec3ee6c14dfd181e44041ce70627c4108f439a58d92 not found: ID does not exist" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.113768 4967 scope.go:117] "RemoveContainer" containerID="2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7" Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.114168 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7\": container with ID starting with 2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7 not found: ID does not exist" containerID="2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.114194 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7"} err="failed to get container status \"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7\": rpc error: code = NotFound desc = could not find container \"2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7\": container with ID starting with 2411f7a5822d652c6338789086e854c206ab8d311f90e05cc94333a114ce1dd7 not found: ID does not exist" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.114537 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/rabbitmq-server-0" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="rabbitmq" containerID="cri-o://68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715" gracePeriod=604800 Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.688026 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.689321 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.690640 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 20 09:25:30 crc kubenswrapper[4967]: E0120 09:25:30.690683 4967 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="keystone-kuttl-tests/openstack-galera-1" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="galera" Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.874177 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:25:30 crc kubenswrapper[4967]: I0120 09:25:30.874375 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" podUID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" containerName="manager" containerID="cri-o://064ee451f731c19f486eac61edbcd28e00e7375cb7bc819851f112a3150b736d" gracePeriod=10 Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.039830 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="keystone-kuttl-tests/memcached-0" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.68:11211: connect: connection refused" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.075470 4967 generic.go:334] "Generic (PLEG): container finished" podID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerID="5a745e943b60dccc0dcba24f62de592b210fe6ab152e694350a3f95b818fccf9" exitCode=0 Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.075561 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"5030695f-a34f-47ef-acaa-c0dfa3601fd8","Type":"ContainerDied","Data":"5a745e943b60dccc0dcba24f62de592b210fe6ab152e694350a3f95b818fccf9"} Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.080238 4967 generic.go:334] "Generic (PLEG): container finished" podID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" containerID="064ee451f731c19f486eac61edbcd28e00e7375cb7bc819851f112a3150b736d" exitCode=0 Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.080274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" event={"ID":"b7bbfe3d-10a4-497f-87e0-d35c154af3ca","Type":"ContainerDied","Data":"064ee451f731c19f486eac61edbcd28e00e7375cb7bc819851f112a3150b736d"} Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.168325 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.168789 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-kxddm" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerName="registry-server" containerID="cri-o://0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" gracePeriod=30 Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.226497 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2"] Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.252880 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/62c62920c746f8aa2c0a4ed190ecdbc8aa8020d6aafcd0adb781e0ce6brsvx2"] Jan 20 09:25:31 crc kubenswrapper[4967]: E0120 09:25:31.518392 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f is running failed: container process not found" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 09:25:31 crc kubenswrapper[4967]: E0120 09:25:31.519420 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f is running failed: container process not found" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 09:25:31 crc kubenswrapper[4967]: E0120 09:25:31.519730 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f is running failed: container process not found" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 20 09:25:31 crc kubenswrapper[4967]: E0120 09:25:31.519782 4967 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f is running failed: container process not found" probeType="Readiness" pod="openstack-operators/keystone-operator-index-kxddm" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerName="registry-server" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.519897 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.626445 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert\") pod \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.626817 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6bpp\" (UniqueName: \"kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp\") pod \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.626874 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert\") pod \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\" (UID: \"b7bbfe3d-10a4-497f-87e0-d35c154af3ca\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.631875 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "b7bbfe3d-10a4-497f-87e0-d35c154af3ca" (UID: "b7bbfe3d-10a4-497f-87e0-d35c154af3ca"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.631926 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "b7bbfe3d-10a4-497f-87e0-d35c154af3ca" (UID: "b7bbfe3d-10a4-497f-87e0-d35c154af3ca"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.633645 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp" (OuterVolumeSpecName: "kube-api-access-m6bpp") pod "b7bbfe3d-10a4-497f-87e0-d35c154af3ca" (UID: "b7bbfe3d-10a4-497f-87e0-d35c154af3ca"). InnerVolumeSpecName "kube-api-access-m6bpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.687306 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.726840 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f730d2-8988-4e09-9cb5-93670237e76a" path="/var/lib/kubelet/pods/06f730d2-8988-4e09-9cb5-93670237e76a/volumes" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.727780 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" path="/var/lib/kubelet/pods/689be023-b3b2-4b2c-b16e-5a24ce36edfc/volumes" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.728345 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqf9m\" (UniqueName: \"kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m\") pod \"01a3a285-14b5-4af7-87de-d3ef4c459794\" (UID: \"01a3a285-14b5-4af7-87de-d3ef4c459794\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.728651 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.728673 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.728684 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6bpp\" (UniqueName: \"kubernetes.io/projected/b7bbfe3d-10a4-497f-87e0-d35c154af3ca-kube-api-access-m6bpp\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.732376 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m" (OuterVolumeSpecName: "kube-api-access-fqf9m") pod "01a3a285-14b5-4af7-87de-d3ef4c459794" (UID: "01a3a285-14b5-4af7-87de-d3ef4c459794"). InnerVolumeSpecName "kube-api-access-fqf9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.804747 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.833328 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqf9m\" (UniqueName: \"kubernetes.io/projected/01a3a285-14b5-4af7-87de-d3ef4c459794-kube-api-access-fqf9m\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.860584 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935051 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ppzp\" (UniqueName: \"kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp\") pod \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935190 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data\") pod \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935222 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935279 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935433 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935471 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935495 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config\") pod \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\" (UID: \"5030695f-a34f-47ef-acaa-c0dfa3601fd8\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935519 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935585 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjmmk\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935638 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd\") pod \"6dcf3472-5ecb-442d-9270-b6da07063df5\" (UID: \"6dcf3472-5ecb-442d-9270-b6da07063df5\") " Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935648 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.935919 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.936164 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data" (OuterVolumeSpecName: "config-data") pod "5030695f-a34f-47ef-acaa-c0dfa3601fd8" (UID: "5030695f-a34f-47ef-acaa-c0dfa3601fd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.936247 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.936325 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5030695f-a34f-47ef-acaa-c0dfa3601fd8" (UID: "5030695f-a34f-47ef-acaa-c0dfa3601fd8"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.936344 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.938829 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp" (OuterVolumeSpecName: "kube-api-access-5ppzp") pod "5030695f-a34f-47ef-acaa-c0dfa3601fd8" (UID: "5030695f-a34f-47ef-acaa-c0dfa3601fd8"). InnerVolumeSpecName "kube-api-access-5ppzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.939378 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk" (OuterVolumeSpecName: "kube-api-access-wjmmk") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "kube-api-access-wjmmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.941072 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info" (OuterVolumeSpecName: "pod-info") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.941179 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.947283 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42" (OuterVolumeSpecName: "persistence") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 20 09:25:31 crc kubenswrapper[4967]: I0120 09:25:31.952361 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.003780 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6dcf3472-5ecb-442d-9270-b6da07063df5" (UID: "6dcf3472-5ecb-442d-9270-b6da07063df5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.019855 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="keystone-kuttl-tests/openstack-galera-0" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="galera" containerID="cri-o://4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0" gracePeriod=26 Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.036892 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.036969 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.036999 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037041 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037100 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037141 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2895c\" (UniqueName: \"kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c\") pod \"7ce51cd1-0945-4880-8571-3779909444d6\" (UID: \"7ce51cd1-0945-4880-8571-3779909444d6\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037377 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ppzp\" (UniqueName: \"kubernetes.io/projected/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kube-api-access-5ppzp\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037396 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037434 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-config-data\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037444 4967 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6dcf3472-5ecb-442d-9270-b6da07063df5-pod-info\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037469 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") on node \"crc\" " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037479 4967 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6dcf3472-5ecb-442d-9270-b6da07063df5-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037488 4967 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5030695f-a34f-47ef-acaa-c0dfa3601fd8-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037496 4967 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6dcf3472-5ecb-442d-9270-b6da07063df5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037504 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjmmk\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-kube-api-access-wjmmk\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037513 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6dcf3472-5ecb-442d-9270-b6da07063df5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037680 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037695 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037886 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.037976 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.040425 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c" (OuterVolumeSpecName: "kube-api-access-2895c") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "kube-api-access-2895c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.046874 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "7ce51cd1-0945-4880-8571-3779909444d6" (UID: "7ce51cd1-0945-4880-8571-3779909444d6"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.050457 4967 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.050592 4967 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42") on node "crc" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.088626 4967 generic.go:334] "Generic (PLEG): container finished" podID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerID="68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715" exitCode=0 Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.088683 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/rabbitmq-server-0" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.088775 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerDied","Data":"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.088812 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/rabbitmq-server-0" event={"ID":"6dcf3472-5ecb-442d-9270-b6da07063df5","Type":"ContainerDied","Data":"150862a26ff3a1ce99025897228e2d43625325b267eaf27152476d8d9f9465e8"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.088835 4967 scope.go:117] "RemoveContainer" containerID="68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.090703 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" event={"ID":"b7bbfe3d-10a4-497f-87e0-d35c154af3ca","Type":"ContainerDied","Data":"199e8252c9af244d80edbb1e24e424f1406aee7a9c079290b09f44916e2a2900"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.090778 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.093737 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/memcached-0" event={"ID":"5030695f-a34f-47ef-acaa-c0dfa3601fd8","Type":"ContainerDied","Data":"1a1599c9a7c37ad6a43b76420cd8dba7b91204b3fa338df24093b5ddd762aa13"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.093798 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/memcached-0" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.095644 4967 generic.go:334] "Generic (PLEG): container finished" podID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" exitCode=0 Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.095710 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kxddm" event={"ID":"01a3a285-14b5-4af7-87de-d3ef4c459794","Type":"ContainerDied","Data":"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.095736 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kxddm" event={"ID":"01a3a285-14b5-4af7-87de-d3ef4c459794","Type":"ContainerDied","Data":"7cbd696484a3c1bb870ca33bb87b5aa78c0828013d7ae3f13c9c96bcd1b663ae"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.095784 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kxddm" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.104104 4967 generic.go:334] "Generic (PLEG): container finished" podID="7ce51cd1-0945-4880-8571-3779909444d6" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" exitCode=0 Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.104145 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerDied","Data":"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.104169 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-1" event={"ID":"7ce51cd1-0945-4880-8571-3779909444d6","Type":"ContainerDied","Data":"0a9eeec82f86eb69bfd3bdd261a62bc1f82e89e2f24d4118d7f080b8d358f4aa"} Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.104225 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-1" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.118947 4967 scope.go:117] "RemoveContainer" containerID="2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138665 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7ce51cd1-0945-4880-8571-3779909444d6-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138697 4967 reconciler_common.go:293] "Volume detached for volume \"pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba3739ff-7786-4e27-9a7f-301cd274ca42\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138712 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2895c\" (UniqueName: \"kubernetes.io/projected/7ce51cd1-0945-4880-8571-3779909444d6-kube-api-access-2895c\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138750 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138765 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138777 4967 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.138787 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7ce51cd1-0945-4880-8571-3779909444d6-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.146743 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.153501 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-74884b4b57-zjgsp"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.154095 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.160672 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.161942 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/rabbitmq-server-0"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.166730 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.170499 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-kxddm"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.177143 4967 scope.go:117] "RemoveContainer" containerID="68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715" Jan 20 09:25:32 crc kubenswrapper[4967]: E0120 09:25:32.178319 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715\": container with ID starting with 68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715 not found: ID does not exist" containerID="68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.178367 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715"} err="failed to get container status \"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715\": rpc error: code = NotFound desc = could not find container \"68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715\": container with ID starting with 68f90f3da5d502c6ed02e016587eecf7c8d963b1e88739e777e7774a467e7715 not found: ID does not exist" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.178397 4967 scope.go:117] "RemoveContainer" containerID="2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.180080 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:25:32 crc kubenswrapper[4967]: E0120 09:25:32.180800 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90\": container with ID starting with 2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90 not found: ID does not exist" containerID="2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.180848 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90"} err="failed to get container status \"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90\": rpc error: code = NotFound desc = could not find container \"2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90\": container with ID starting with 2bf7f83f57fd0bc896807a286176172a98c3b9f987c7f62c65a47ca35886cf90 not found: ID does not exist" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.180880 4967 scope.go:117] "RemoveContainer" containerID="064ee451f731c19f486eac61edbcd28e00e7375cb7bc819851f112a3150b736d" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.191996 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/memcached-0"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.198820 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.203342 4967 scope.go:117] "RemoveContainer" containerID="5a745e943b60dccc0dcba24f62de592b210fe6ab152e694350a3f95b818fccf9" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.204975 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-1"] Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.219416 4967 scope.go:117] "RemoveContainer" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.234362 4967 scope.go:117] "RemoveContainer" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" Jan 20 09:25:32 crc kubenswrapper[4967]: E0120 09:25:32.235021 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f\": container with ID starting with 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f not found: ID does not exist" containerID="0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.235083 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f"} err="failed to get container status \"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f\": rpc error: code = NotFound desc = could not find container \"0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f\": container with ID starting with 0c2ef61c069a91019ebd77f20dcbd7eac37cc6068e7720b664f0d3f04d91cc8f not found: ID does not exist" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.235117 4967 scope.go:117] "RemoveContainer" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.239548 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.251845 4967 scope.go:117] "RemoveContainer" containerID="598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.267954 4967 scope.go:117] "RemoveContainer" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" Jan 20 09:25:32 crc kubenswrapper[4967]: E0120 09:25:32.268384 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d\": container with ID starting with 75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d not found: ID does not exist" containerID="75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.268464 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d"} err="failed to get container status \"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d\": rpc error: code = NotFound desc = could not find container \"75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d\": container with ID starting with 75d286f8bc542e2ed53ff0f58509757378bc8b78559ddc3c8be0b433f5372a5d not found: ID does not exist" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.268499 4967 scope.go:117] "RemoveContainer" containerID="598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9" Jan 20 09:25:32 crc kubenswrapper[4967]: E0120 09:25:32.268832 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9\": container with ID starting with 598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9 not found: ID does not exist" containerID="598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.268863 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9"} err="failed to get container status \"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9\": rpc error: code = NotFound desc = could not find container \"598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9\": container with ID starting with 598b6eced3d1406a95c4b1843de199cb8e5ccadcd93d2da671ce761dcc5940c9 not found: ID does not exist" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.782353 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845203 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845247 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xshsd\" (UniqueName: \"kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845317 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845340 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845394 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\" (UID: \"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22\") " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.845863 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.846058 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.846206 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.848458 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd" (OuterVolumeSpecName: "kube-api-access-xshsd") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "kube-api-access-xshsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.853509 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" (UID: "cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.946964 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.947023 4967 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.947046 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xshsd\" (UniqueName: \"kubernetes.io/projected/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-kube-api-access-xshsd\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.947062 4967 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.947074 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.947086 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:32 crc kubenswrapper[4967]: I0120 09:25:32.956385 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.048756 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.111868 4967 generic.go:334] "Generic (PLEG): container finished" podID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerID="4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0" exitCode=0 Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.111932 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keystone-kuttl-tests/openstack-galera-0" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.111968 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerDied","Data":"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0"} Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.112032 4967 scope.go:117] "RemoveContainer" containerID="4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.112021 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="keystone-kuttl-tests/openstack-galera-0" event={"ID":"cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22","Type":"ContainerDied","Data":"7ea6cf633ff8f5d66a0bb2857d37b1efbb3f36a46e7c2e67570616c2913482ea"} Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.133850 4967 scope.go:117] "RemoveContainer" containerID="8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.153109 4967 scope.go:117] "RemoveContainer" containerID="4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0" Jan 20 09:25:33 crc kubenswrapper[4967]: E0120 09:25:33.153518 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0\": container with ID starting with 4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0 not found: ID does not exist" containerID="4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.153562 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0"} err="failed to get container status \"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0\": rpc error: code = NotFound desc = could not find container \"4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0\": container with ID starting with 4df447b21d83b028d8dc2e9c84926c9dcea1ca24f50398940079c80e375eede0 not found: ID does not exist" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.153590 4967 scope.go:117] "RemoveContainer" containerID="8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0" Jan 20 09:25:33 crc kubenswrapper[4967]: E0120 09:25:33.153975 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0\": container with ID starting with 8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0 not found: ID does not exist" containerID="8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.154013 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0"} err="failed to get container status \"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0\": rpc error: code = NotFound desc = could not find container \"8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0\": container with ID starting with 8d64bd68e3fe9008d4b5789fe1598c0c094bdc06fdc1eba2a34e09f1075623e0 not found: ID does not exist" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.160104 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.166224 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["keystone-kuttl-tests/openstack-galera-0"] Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.705265 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" path="/var/lib/kubelet/pods/01a3a285-14b5-4af7-87de-d3ef4c459794/volumes" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.706003 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" path="/var/lib/kubelet/pods/5030695f-a34f-47ef-acaa-c0dfa3601fd8/volumes" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.706985 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" path="/var/lib/kubelet/pods/6dcf3472-5ecb-442d-9270-b6da07063df5/volumes" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.708275 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce51cd1-0945-4880-8571-3779909444d6" path="/var/lib/kubelet/pods/7ce51cd1-0945-4880-8571-3779909444d6/volumes" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.708942 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" path="/var/lib/kubelet/pods/b7bbfe3d-10a4-497f-87e0-d35c154af3ca/volumes" Jan 20 09:25:33 crc kubenswrapper[4967]: I0120 09:25:33.709549 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" path="/var/lib/kubelet/pods/cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22/volumes" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059197 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059795 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059810 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059824 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059832 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059844 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerName="memcached" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059852 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerName="memcached" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059862 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" containerName="manager" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059869 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" containerName="manager" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059890 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="setup-container" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059897 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="setup-container" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059909 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerName="registry-server" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059916 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerName="registry-server" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059927 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059934 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059945 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="rabbitmq" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059952 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="rabbitmq" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059964 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059971 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.059989 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.059996 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="mysql-bootstrap" Jan 20 09:25:38 crc kubenswrapper[4967]: E0120 09:25:38.060005 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060013 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060149 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a3a285-14b5-4af7-87de-d3ef4c459794" containerName="registry-server" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060168 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5030695f-a34f-47ef-acaa-c0dfa3601fd8" containerName="memcached" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060180 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="689be023-b3b2-4b2c-b16e-5a24ce36edfc" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060193 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd3d19e-4ca0-4a64-9a2d-a7dd13f8be22" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060201 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7bbfe3d-10a4-497f-87e0-d35c154af3ca" containerName="manager" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060212 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce51cd1-0945-4880-8571-3779909444d6" containerName="galera" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.060223 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dcf3472-5ecb-442d-9270-b6da07063df5" containerName="rabbitmq" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.061175 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.075127 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.116464 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6lqp\" (UniqueName: \"kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.116530 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.116593 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.200712 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.201229 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" podUID="a4fc08d9-6214-48db-9f0c-38cda7310168" containerName="manager" containerID="cri-o://8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c" gracePeriod=10 Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.217812 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.217969 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6lqp\" (UniqueName: \"kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.218096 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.218665 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.219774 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.250328 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6lqp\" (UniqueName: \"kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp\") pod \"community-operators-v4fz4\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.382854 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.484822 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.485059 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-4b7zw" podUID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" containerName="registry-server" containerID="cri-o://159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f" gracePeriod=30 Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.509597 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.518194 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/ea82580bc5724477f94b47db468c840840d4aaf95efc52f7d04b6353c1rdlj9"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.727347 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.738942 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.824467 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp8tc\" (UniqueName: \"kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc\") pod \"a4fc08d9-6214-48db-9f0c-38cda7310168\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.824513 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert\") pod \"a4fc08d9-6214-48db-9f0c-38cda7310168\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.824539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert\") pod \"a4fc08d9-6214-48db-9f0c-38cda7310168\" (UID: \"a4fc08d9-6214-48db-9f0c-38cda7310168\") " Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.831048 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "a4fc08d9-6214-48db-9f0c-38cda7310168" (UID: "a4fc08d9-6214-48db-9f0c-38cda7310168"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.831120 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "a4fc08d9-6214-48db-9f0c-38cda7310168" (UID: "a4fc08d9-6214-48db-9f0c-38cda7310168"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.832880 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc" (OuterVolumeSpecName: "kube-api-access-rp8tc") pod "a4fc08d9-6214-48db-9f0c-38cda7310168" (UID: "a4fc08d9-6214-48db-9f0c-38cda7310168"). InnerVolumeSpecName "kube-api-access-rp8tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.926358 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.926384 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp8tc\" (UniqueName: \"kubernetes.io/projected/a4fc08d9-6214-48db-9f0c-38cda7310168-kube-api-access-rp8tc\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:38 crc kubenswrapper[4967]: I0120 09:25:38.926394 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4fc08d9-6214-48db-9f0c-38cda7310168-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.029954 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.128177 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc9nr\" (UniqueName: \"kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr\") pod \"a1eac7ed-6381-4765-b9a6-1b680641c4c0\" (UID: \"a1eac7ed-6381-4765-b9a6-1b680641c4c0\") " Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.134726 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr" (OuterVolumeSpecName: "kube-api-access-vc9nr") pod "a1eac7ed-6381-4765-b9a6-1b680641c4c0" (UID: "a1eac7ed-6381-4765-b9a6-1b680641c4c0"). InnerVolumeSpecName "kube-api-access-vc9nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.153702 4967 generic.go:334] "Generic (PLEG): container finished" podID="a4fc08d9-6214-48db-9f0c-38cda7310168" containerID="8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c" exitCode=0 Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.153760 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" event={"ID":"a4fc08d9-6214-48db-9f0c-38cda7310168","Type":"ContainerDied","Data":"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.153820 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" event={"ID":"a4fc08d9-6214-48db-9f0c-38cda7310168","Type":"ContainerDied","Data":"45f10f613d721e069edb97db9c86894cee5c69028143fa77ca0d7c3720f70296"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.153776 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.153848 4967 scope.go:117] "RemoveContainer" containerID="8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.155804 4967 generic.go:334] "Generic (PLEG): container finished" podID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerID="6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03" exitCode=0 Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.155896 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerDied","Data":"6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.155923 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerStarted","Data":"d7c7a2682a7c00549911068b9750fdca5288ab6b263282aec4df8ba7a3203bee"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.158007 4967 generic.go:334] "Generic (PLEG): container finished" podID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" containerID="159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f" exitCode=0 Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.158043 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4b7zw" event={"ID":"a1eac7ed-6381-4765-b9a6-1b680641c4c0","Type":"ContainerDied","Data":"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.158051 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-4b7zw" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.158077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-4b7zw" event={"ID":"a1eac7ed-6381-4765-b9a6-1b680641c4c0","Type":"ContainerDied","Data":"01fcbed6d7ae8076502bcfb6c4ee28c040467185f0010b43a3b46caebfc06fc3"} Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.177096 4967 scope.go:117] "RemoveContainer" containerID="8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c" Jan 20 09:25:39 crc kubenswrapper[4967]: E0120 09:25:39.177586 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c\": container with ID starting with 8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c not found: ID does not exist" containerID="8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.177662 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c"} err="failed to get container status \"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c\": rpc error: code = NotFound desc = could not find container \"8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c\": container with ID starting with 8b03b81fff5fd61dd24823df6829780929f3408a853b8056b4d93521416efe2c not found: ID does not exist" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.177691 4967 scope.go:117] "RemoveContainer" containerID="159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.197384 4967 scope.go:117] "RemoveContainer" containerID="159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f" Jan 20 09:25:39 crc kubenswrapper[4967]: E0120 09:25:39.197952 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f\": container with ID starting with 159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f not found: ID does not exist" containerID="159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.198001 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f"} err="failed to get container status \"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f\": rpc error: code = NotFound desc = could not find container \"159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f\": container with ID starting with 159b45efb8c12dbfc9bd8f1683e2563c5379cc1581e9aff9ecaeb853acd4fb9f not found: ID does not exist" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.202926 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.212274 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-4b7zw"] Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.217302 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.223352 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6f9888886b-cbv7g"] Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.230270 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc9nr\" (UniqueName: \"kubernetes.io/projected/a1eac7ed-6381-4765-b9a6-1b680641c4c0-kube-api-access-vc9nr\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.702599 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" path="/var/lib/kubelet/pods/a1eac7ed-6381-4765-b9a6-1b680641c4c0/volumes" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.703574 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4fc08d9-6214-48db-9f0c-38cda7310168" path="/var/lib/kubelet/pods/a4fc08d9-6214-48db-9f0c-38cda7310168/volumes" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.704711 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d010afb9-5722-465f-a06d-bd3edaacc210" path="/var/lib/kubelet/pods/d010afb9-5722-465f-a06d-bd3edaacc210/volumes" Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.824150 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:25:39 crc kubenswrapper[4967]: I0120 09:25:39.824398 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" containerID="cri-o://706826535e4a9cc773ec188a6496d0f3c6ced4eca112d3293f987ea14d53a4f6" gracePeriod=10 Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.115640 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.116037 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-lk86j" podUID="45b78b67-e141-49df-ba8f-6e8f7958304b" containerName="registry-server" containerID="cri-o://7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561" gracePeriod=30 Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.157405 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l"] Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.165536 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/1db583f94d1987eb997cb67ef7db92b9cbf8f229674982c4f1534d76a9wsq5l"] Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.171033 4967 generic.go:334] "Generic (PLEG): container finished" podID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerID="706826535e4a9cc773ec188a6496d0f3c6ced4eca112d3293f987ea14d53a4f6" exitCode=0 Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.171171 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerDied","Data":"706826535e4a9cc773ec188a6496d0f3c6ced4eca112d3293f987ea14d53a4f6"} Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.171222 4967 scope.go:117] "RemoveContainer" containerID="8946ebfdbb11e9d1b449ffdce1dcc75cdf64bec416e1d16252aaf05ccec4987f" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.180938 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerStarted","Data":"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb"} Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.352586 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.450318 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert\") pod \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.450382 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert\") pod \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.450430 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrhnb\" (UniqueName: \"kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb\") pod \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\" (UID: \"49a21c96-9728-4237-b4a2-ce0d3ac40ef6\") " Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.456890 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "49a21c96-9728-4237-b4a2-ce0d3ac40ef6" (UID: "49a21c96-9728-4237-b4a2-ce0d3ac40ef6"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.457304 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "49a21c96-9728-4237-b4a2-ce0d3ac40ef6" (UID: "49a21c96-9728-4237-b4a2-ce0d3ac40ef6"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.460814 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb" (OuterVolumeSpecName: "kube-api-access-xrhnb") pod "49a21c96-9728-4237-b4a2-ce0d3ac40ef6" (UID: "49a21c96-9728-4237-b4a2-ce0d3ac40ef6"). InnerVolumeSpecName "kube-api-access-xrhnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.546389 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.552638 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.552672 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.552686 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrhnb\" (UniqueName: \"kubernetes.io/projected/49a21c96-9728-4237-b4a2-ce0d3ac40ef6-kube-api-access-xrhnb\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.653546 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8s2f\" (UniqueName: \"kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f\") pod \"45b78b67-e141-49df-ba8f-6e8f7958304b\" (UID: \"45b78b67-e141-49df-ba8f-6e8f7958304b\") " Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.656278 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f" (OuterVolumeSpecName: "kube-api-access-c8s2f") pod "45b78b67-e141-49df-ba8f-6e8f7958304b" (UID: "45b78b67-e141-49df-ba8f-6e8f7958304b"). InnerVolumeSpecName "kube-api-access-c8s2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:40 crc kubenswrapper[4967]: I0120 09:25:40.754680 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8s2f\" (UniqueName: \"kubernetes.io/projected/45b78b67-e141-49df-ba8f-6e8f7958304b-kube-api-access-c8s2f\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.187359 4967 generic.go:334] "Generic (PLEG): container finished" podID="45b78b67-e141-49df-ba8f-6e8f7958304b" containerID="7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561" exitCode=0 Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.187407 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lk86j" event={"ID":"45b78b67-e141-49df-ba8f-6e8f7958304b","Type":"ContainerDied","Data":"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561"} Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.188285 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lk86j" event={"ID":"45b78b67-e141-49df-ba8f-6e8f7958304b","Type":"ContainerDied","Data":"410d7d308a46be25b956ba97bc50717c51a78d7a265d9bf93160278d743f0d66"} Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.187441 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lk86j" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.188314 4967 scope.go:117] "RemoveContainer" containerID="7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.189880 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.191325 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk" event={"ID":"49a21c96-9728-4237-b4a2-ce0d3ac40ef6","Type":"ContainerDied","Data":"380a030fdf312547d4364c2ee0b46d539608cfbd8bb7044fcb9a2637718cdbc8"} Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.192522 4967 generic.go:334] "Generic (PLEG): container finished" podID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerID="00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb" exitCode=0 Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.192645 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerDied","Data":"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb"} Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.204070 4967 scope.go:117] "RemoveContainer" containerID="7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561" Jan 20 09:25:41 crc kubenswrapper[4967]: E0120 09:25:41.204526 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561\": container with ID starting with 7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561 not found: ID does not exist" containerID="7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.204578 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561"} err="failed to get container status \"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561\": rpc error: code = NotFound desc = could not find container \"7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561\": container with ID starting with 7e16e3df6d40c0a035bb71c10086588f0a513821d54a7c2f0be486094817f561 not found: ID does not exist" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.204641 4967 scope.go:117] "RemoveContainer" containerID="706826535e4a9cc773ec188a6496d0f3c6ced4eca112d3293f987ea14d53a4f6" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.390397 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.395289 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7f8fd5d49d-4bblk"] Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.402625 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.407766 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-lk86j"] Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.703291 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45b78b67-e141-49df-ba8f-6e8f7958304b" path="/var/lib/kubelet/pods/45b78b67-e141-49df-ba8f-6e8f7958304b/volumes" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.704319 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" path="/var/lib/kubelet/pods/49a21c96-9728-4237-b4a2-ce0d3ac40ef6/volumes" Jan 20 09:25:41 crc kubenswrapper[4967]: I0120 09:25:41.705306 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983b1332-9808-49d1-a25d-2a76ee5f9ce8" path="/var/lib/kubelet/pods/983b1332-9808-49d1-a25d-2a76ee5f9ce8/volumes" Jan 20 09:25:42 crc kubenswrapper[4967]: I0120 09:25:42.200459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerStarted","Data":"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86"} Jan 20 09:25:42 crc kubenswrapper[4967]: I0120 09:25:42.223528 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v4fz4" podStartSLOduration=1.368001129 podStartE2EDuration="4.223507321s" podCreationTimestamp="2026-01-20 09:25:38 +0000 UTC" firstStartedPulling="2026-01-20 09:25:39.157458368 +0000 UTC m=+2173.742758575" lastFinishedPulling="2026-01-20 09:25:42.01296455 +0000 UTC m=+2176.598264767" observedRunningTime="2026-01-20 09:25:42.219113641 +0000 UTC m=+2176.804413858" watchObservedRunningTime="2026-01-20 09:25:42.223507321 +0000 UTC m=+2176.808807528" Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.451703 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.451901 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" podUID="40c1982c-ddc0-4043-a956-2a5332612176" containerName="operator" containerID="cri-o://6b4074506c651043bcbf53da34b1aa44d340376b57f491fe9a75004cd975fc1d" gracePeriod=10 Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.860230 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.860474 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" podUID="82afd249-e563-437f-9011-05be3f663020" containerName="registry-server" containerID="cri-o://2eec8faabd74788596b928356698a73692a061101b4c0e23b10f2ad38c549cde" gracePeriod=30 Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.896300 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8"] Jan 20 09:25:43 crc kubenswrapper[4967]: I0120 09:25:43.899179 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5904zkw8"] Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.221922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" event={"ID":"40c1982c-ddc0-4043-a956-2a5332612176","Type":"ContainerDied","Data":"6b4074506c651043bcbf53da34b1aa44d340376b57f491fe9a75004cd975fc1d"} Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.222755 4967 generic.go:334] "Generic (PLEG): container finished" podID="40c1982c-ddc0-4043-a956-2a5332612176" containerID="6b4074506c651043bcbf53da34b1aa44d340376b57f491fe9a75004cd975fc1d" exitCode=0 Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.233395 4967 generic.go:334] "Generic (PLEG): container finished" podID="82afd249-e563-437f-9011-05be3f663020" containerID="2eec8faabd74788596b928356698a73692a061101b4c0e23b10f2ad38c549cde" exitCode=0 Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.233462 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" event={"ID":"82afd249-e563-437f-9011-05be3f663020","Type":"ContainerDied","Data":"2eec8faabd74788596b928356698a73692a061101b4c0e23b10f2ad38c549cde"} Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.295109 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.370174 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.398494 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lwlr\" (UniqueName: \"kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr\") pod \"40c1982c-ddc0-4043-a956-2a5332612176\" (UID: \"40c1982c-ddc0-4043-a956-2a5332612176\") " Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.398557 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxkpj\" (UniqueName: \"kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj\") pod \"82afd249-e563-437f-9011-05be3f663020\" (UID: \"82afd249-e563-437f-9011-05be3f663020\") " Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.404540 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj" (OuterVolumeSpecName: "kube-api-access-hxkpj") pod "82afd249-e563-437f-9011-05be3f663020" (UID: "82afd249-e563-437f-9011-05be3f663020"). InnerVolumeSpecName "kube-api-access-hxkpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.404859 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr" (OuterVolumeSpecName: "kube-api-access-4lwlr") pod "40c1982c-ddc0-4043-a956-2a5332612176" (UID: "40c1982c-ddc0-4043-a956-2a5332612176"). InnerVolumeSpecName "kube-api-access-4lwlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.500162 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lwlr\" (UniqueName: \"kubernetes.io/projected/40c1982c-ddc0-4043-a956-2a5332612176-kube-api-access-4lwlr\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:44 crc kubenswrapper[4967]: I0120 09:25:44.500193 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxkpj\" (UniqueName: \"kubernetes.io/projected/82afd249-e563-437f-9011-05be3f663020-kube-api-access-hxkpj\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.241238 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" event={"ID":"82afd249-e563-437f-9011-05be3f663020","Type":"ContainerDied","Data":"fa99a32ca18dbe687ccddac6930948876d537078aac2d861c07b95b832bf11d3"} Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.241281 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-qsnbm" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.241525 4967 scope.go:117] "RemoveContainer" containerID="2eec8faabd74788596b928356698a73692a061101b4c0e23b10f2ad38c549cde" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.243331 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" event={"ID":"40c1982c-ddc0-4043-a956-2a5332612176","Type":"ContainerDied","Data":"af6937c7704c79a09486dfd57b27752ee614d0620e6d29903d8ec0e088402dcb"} Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.243482 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.263442 4967 scope.go:117] "RemoveContainer" containerID="6b4074506c651043bcbf53da34b1aa44d340376b57f491fe9a75004cd975fc1d" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.286947 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:25:45 crc kubenswrapper[4967]: E0120 09:25:45.293769 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82afd249_e563_437f_9011_05be3f663020.slice/crio-fa99a32ca18dbe687ccddac6930948876d537078aac2d861c07b95b832bf11d3\": RecentStats: unable to find data in memory cache]" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.296323 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzkln"] Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.300091 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.303106 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-qsnbm"] Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.701277 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c1982c-ddc0-4043-a956-2a5332612176" path="/var/lib/kubelet/pods/40c1982c-ddc0-4043-a956-2a5332612176/volumes" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.702277 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82afd249-e563-437f-9011-05be3f663020" path="/var/lib/kubelet/pods/82afd249-e563-437f-9011-05be3f663020/volumes" Jan 20 09:25:45 crc kubenswrapper[4967]: I0120 09:25:45.703148 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2" path="/var/lib/kubelet/pods/cbd0e9ab-c6a2-4db9-80bb-06b60488c0d2/volumes" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.662791 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663416 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fc08d9-6214-48db-9f0c-38cda7310168" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663437 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fc08d9-6214-48db-9f0c-38cda7310168" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663457 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82afd249-e563-437f-9011-05be3f663020" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663469 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="82afd249-e563-437f-9011-05be3f663020" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663483 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b78b67-e141-49df-ba8f-6e8f7958304b" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663498 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b78b67-e141-49df-ba8f-6e8f7958304b" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663517 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663529 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663553 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663564 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.663585 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c1982c-ddc0-4043-a956-2a5332612176" containerName="operator" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663596 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c1982c-ddc0-4043-a956-2a5332612176" containerName="operator" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663835 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663859 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663872 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="82afd249-e563-437f-9011-05be3f663020" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663888 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c1982c-ddc0-4043-a956-2a5332612176" containerName="operator" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663902 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fc08d9-6214-48db-9f0c-38cda7310168" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663926 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1eac7ed-6381-4765-b9a6-1b680641c4c0" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.663948 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="45b78b67-e141-49df-ba8f-6e8f7958304b" containerName="registry-server" Jan 20 09:25:47 crc kubenswrapper[4967]: E0120 09:25:47.664127 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.664142 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a21c96-9728-4237-b4a2-ce0d3ac40ef6" containerName="manager" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.665647 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.669427 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.739769 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.739852 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n92ft\" (UniqueName: \"kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.739880 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.841820 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.842172 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n92ft\" (UniqueName: \"kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.842305 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.842417 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.842819 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.862998 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n92ft\" (UniqueName: \"kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft\") pod \"redhat-operators-zmlcg\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:47 crc kubenswrapper[4967]: I0120 09:25:47.991074 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.383315 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.383584 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.433765 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.459419 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:25:48 crc kubenswrapper[4967]: W0120 09:25:48.471582 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod913ab81d_ce17_4e90_b6cd_7072178ee4ef.slice/crio-b24b3b188259c389e2e0b6790884c76409c46b1f69c1a1cdd3e77cd57ce41876 WatchSource:0}: Error finding container b24b3b188259c389e2e0b6790884c76409c46b1f69c1a1cdd3e77cd57ce41876: Status 404 returned error can't find the container with id b24b3b188259c389e2e0b6790884c76409c46b1f69c1a1cdd3e77cd57ce41876 Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.656519 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.658108 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.670748 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.755564 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.755644 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.755669 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnlh\" (UniqueName: \"kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.857258 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.857314 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnlh\" (UniqueName: \"kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.857402 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.857842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.857886 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.881473 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnlh\" (UniqueName: \"kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh\") pod \"redhat-marketplace-9zfz4\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:48 crc kubenswrapper[4967]: I0120 09:25:48.971252 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:49 crc kubenswrapper[4967]: I0120 09:25:49.273127 4967 generic.go:334] "Generic (PLEG): container finished" podID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerID="9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db" exitCode=0 Jan 20 09:25:49 crc kubenswrapper[4967]: I0120 09:25:49.273175 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerDied","Data":"9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db"} Jan 20 09:25:49 crc kubenswrapper[4967]: I0120 09:25:49.274295 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerStarted","Data":"b24b3b188259c389e2e0b6790884c76409c46b1f69c1a1cdd3e77cd57ce41876"} Jan 20 09:25:49 crc kubenswrapper[4967]: I0120 09:25:49.320690 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:49 crc kubenswrapper[4967]: I0120 09:25:49.374952 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:25:50 crc kubenswrapper[4967]: I0120 09:25:50.282208 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerStarted","Data":"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516"} Jan 20 09:25:50 crc kubenswrapper[4967]: I0120 09:25:50.283677 4967 generic.go:334] "Generic (PLEG): container finished" podID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerID="8809ab31a7c73679cb9e02aecbb88d495663f0424b693a4b9ec0102a0b84246d" exitCode=0 Jan 20 09:25:50 crc kubenswrapper[4967]: I0120 09:25:50.283777 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerDied","Data":"8809ab31a7c73679cb9e02aecbb88d495663f0424b693a4b9ec0102a0b84246d"} Jan 20 09:25:50 crc kubenswrapper[4967]: I0120 09:25:50.283812 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerStarted","Data":"163b79cc1a3b693258202b9ca664eb663e51d020645bc13a5e7f5dfb3938c2e2"} Jan 20 09:25:51 crc kubenswrapper[4967]: I0120 09:25:51.054447 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:51 crc kubenswrapper[4967]: I0120 09:25:51.294639 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v4fz4" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="registry-server" containerID="cri-o://9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86" gracePeriod=2 Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.193741 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.301350 4967 generic.go:334] "Generic (PLEG): container finished" podID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerID="80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516" exitCode=0 Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.301473 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerDied","Data":"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516"} Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.302892 4967 generic.go:334] "Generic (PLEG): container finished" podID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerID="322307c020d73e2a592a000c88fa36f175c5664b460775ae1280cd63710b4cb0" exitCode=0 Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.302920 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerDied","Data":"322307c020d73e2a592a000c88fa36f175c5664b460775ae1280cd63710b4cb0"} Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.306715 4967 generic.go:334] "Generic (PLEG): container finished" podID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerID="9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86" exitCode=0 Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.306754 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerDied","Data":"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86"} Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.306799 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v4fz4" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.306824 4967 scope.go:117] "RemoveContainer" containerID="9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.306812 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v4fz4" event={"ID":"1c46e441-3dde-49ed-b75c-38ce4483948b","Type":"ContainerDied","Data":"d7c7a2682a7c00549911068b9750fdca5288ab6b263282aec4df8ba7a3203bee"} Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.322519 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6lqp\" (UniqueName: \"kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp\") pod \"1c46e441-3dde-49ed-b75c-38ce4483948b\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.322676 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content\") pod \"1c46e441-3dde-49ed-b75c-38ce4483948b\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.322720 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities\") pod \"1c46e441-3dde-49ed-b75c-38ce4483948b\" (UID: \"1c46e441-3dde-49ed-b75c-38ce4483948b\") " Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.323854 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities" (OuterVolumeSpecName: "utilities") pod "1c46e441-3dde-49ed-b75c-38ce4483948b" (UID: "1c46e441-3dde-49ed-b75c-38ce4483948b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.341884 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp" (OuterVolumeSpecName: "kube-api-access-c6lqp") pod "1c46e441-3dde-49ed-b75c-38ce4483948b" (UID: "1c46e441-3dde-49ed-b75c-38ce4483948b"). InnerVolumeSpecName "kube-api-access-c6lqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.373789 4967 scope.go:117] "RemoveContainer" containerID="00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.388529 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c46e441-3dde-49ed-b75c-38ce4483948b" (UID: "1c46e441-3dde-49ed-b75c-38ce4483948b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.428318 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.428350 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6lqp\" (UniqueName: \"kubernetes.io/projected/1c46e441-3dde-49ed-b75c-38ce4483948b-kube-api-access-c6lqp\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.428361 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c46e441-3dde-49ed-b75c-38ce4483948b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.437311 4967 scope.go:117] "RemoveContainer" containerID="6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.478047 4967 scope.go:117] "RemoveContainer" containerID="9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86" Jan 20 09:25:52 crc kubenswrapper[4967]: E0120 09:25:52.478543 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86\": container with ID starting with 9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86 not found: ID does not exist" containerID="9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.478593 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86"} err="failed to get container status \"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86\": rpc error: code = NotFound desc = could not find container \"9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86\": container with ID starting with 9aaacc10f0b7ddbf91ad8fc95ddcb73a94ea3503cdaa5127f7671ddfcb4b4b86 not found: ID does not exist" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.478637 4967 scope.go:117] "RemoveContainer" containerID="00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb" Jan 20 09:25:52 crc kubenswrapper[4967]: E0120 09:25:52.478913 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb\": container with ID starting with 00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb not found: ID does not exist" containerID="00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.478940 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb"} err="failed to get container status \"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb\": rpc error: code = NotFound desc = could not find container \"00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb\": container with ID starting with 00fb2f1dd1e44b8577ca56e45d1f0e9d0f3e4c4b45da8abe9eee3465be3ad6fb not found: ID does not exist" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.478960 4967 scope.go:117] "RemoveContainer" containerID="6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03" Jan 20 09:25:52 crc kubenswrapper[4967]: E0120 09:25:52.479149 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03\": container with ID starting with 6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03 not found: ID does not exist" containerID="6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.479169 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03"} err="failed to get container status \"6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03\": rpc error: code = NotFound desc = could not find container \"6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03\": container with ID starting with 6b228dcb556fdc247173fb658251f262064c2c95b194b2d3d200034c1776ec03 not found: ID does not exist" Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.650502 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:52 crc kubenswrapper[4967]: I0120 09:25:52.667557 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v4fz4"] Jan 20 09:25:53 crc kubenswrapper[4967]: I0120 09:25:53.318336 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerStarted","Data":"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2"} Jan 20 09:25:53 crc kubenswrapper[4967]: I0120 09:25:53.321684 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerStarted","Data":"461d4b7cbe26c5345eda98a21997343f527911c3b15a2c50a6bbddc1ca705ec8"} Jan 20 09:25:53 crc kubenswrapper[4967]: I0120 09:25:53.338552 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zmlcg" podStartSLOduration=2.819537637 podStartE2EDuration="6.338531799s" podCreationTimestamp="2026-01-20 09:25:47 +0000 UTC" firstStartedPulling="2026-01-20 09:25:49.27450881 +0000 UTC m=+2183.859809017" lastFinishedPulling="2026-01-20 09:25:52.793502972 +0000 UTC m=+2187.378803179" observedRunningTime="2026-01-20 09:25:53.335527488 +0000 UTC m=+2187.920827695" watchObservedRunningTime="2026-01-20 09:25:53.338531799 +0000 UTC m=+2187.923832006" Jan 20 09:25:53 crc kubenswrapper[4967]: I0120 09:25:53.360313 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9zfz4" podStartSLOduration=2.674628164 podStartE2EDuration="5.360281171s" podCreationTimestamp="2026-01-20 09:25:48 +0000 UTC" firstStartedPulling="2026-01-20 09:25:50.28547748 +0000 UTC m=+2184.870777687" lastFinishedPulling="2026-01-20 09:25:52.971130487 +0000 UTC m=+2187.556430694" observedRunningTime="2026-01-20 09:25:53.354033541 +0000 UTC m=+2187.939333758" watchObservedRunningTime="2026-01-20 09:25:53.360281171 +0000 UTC m=+2187.945581418" Jan 20 09:25:53 crc kubenswrapper[4967]: I0120 09:25:53.700553 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" path="/var/lib/kubelet/pods/1c46e441-3dde-49ed-b75c-38ce4483948b/volumes" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.287496 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wp7vg/must-gather-r4fvk"] Jan 20 09:25:57 crc kubenswrapper[4967]: E0120 09:25:57.288076 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="registry-server" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.288091 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="registry-server" Jan 20 09:25:57 crc kubenswrapper[4967]: E0120 09:25:57.288113 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="extract-content" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.288120 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="extract-content" Jan 20 09:25:57 crc kubenswrapper[4967]: E0120 09:25:57.288142 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="extract-utilities" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.288151 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="extract-utilities" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.288269 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c46e441-3dde-49ed-b75c-38ce4483948b" containerName="registry-server" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.288990 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.290995 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wp7vg"/"default-dockercfg-ph9nk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.291553 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wp7vg"/"openshift-service-ca.crt" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.291562 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wp7vg"/"kube-root-ca.crt" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.304782 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wp7vg/must-gather-r4fvk"] Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.388935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.389039 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z78r4\" (UniqueName: \"kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.490507 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z78r4\" (UniqueName: \"kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.490584 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.491021 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.508734 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z78r4\" (UniqueName: \"kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4\") pod \"must-gather-r4fvk\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.606953 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.991916 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:57 crc kubenswrapper[4967]: I0120 09:25:57.992856 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:25:58 crc kubenswrapper[4967]: I0120 09:25:58.021959 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wp7vg/must-gather-r4fvk"] Jan 20 09:25:58 crc kubenswrapper[4967]: W0120 09:25:58.030572 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561359f4_e525_49a0_b560_d63c5491987b.slice/crio-5482f49961a7b57897b9db0a96e4acb7c5fee7a6be80f5806e7f333af8041e4a WatchSource:0}: Error finding container 5482f49961a7b57897b9db0a96e4acb7c5fee7a6be80f5806e7f333af8041e4a: Status 404 returned error can't find the container with id 5482f49961a7b57897b9db0a96e4acb7c5fee7a6be80f5806e7f333af8041e4a Jan 20 09:25:58 crc kubenswrapper[4967]: I0120 09:25:58.358010 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" event={"ID":"561359f4-e525-49a0-b560-d63c5491987b","Type":"ContainerStarted","Data":"5482f49961a7b57897b9db0a96e4acb7c5fee7a6be80f5806e7f333af8041e4a"} Jan 20 09:25:58 crc kubenswrapper[4967]: I0120 09:25:58.971396 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:58 crc kubenswrapper[4967]: I0120 09:25:58.971441 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:59 crc kubenswrapper[4967]: I0120 09:25:59.009200 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:59 crc kubenswrapper[4967]: I0120 09:25:59.039745 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zmlcg" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="registry-server" probeResult="failure" output=< Jan 20 09:25:59 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Jan 20 09:25:59 crc kubenswrapper[4967]: > Jan 20 09:25:59 crc kubenswrapper[4967]: I0120 09:25:59.415503 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:25:59 crc kubenswrapper[4967]: I0120 09:25:59.648864 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:26:01 crc kubenswrapper[4967]: I0120 09:26:01.379035 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9zfz4" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="registry-server" containerID="cri-o://461d4b7cbe26c5345eda98a21997343f527911c3b15a2c50a6bbddc1ca705ec8" gracePeriod=2 Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.405236 4967 generic.go:334] "Generic (PLEG): container finished" podID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerID="461d4b7cbe26c5345eda98a21997343f527911c3b15a2c50a6bbddc1ca705ec8" exitCode=0 Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.405318 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerDied","Data":"461d4b7cbe26c5345eda98a21997343f527911c3b15a2c50a6bbddc1ca705ec8"} Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.539980 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.678597 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content\") pod \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.678695 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities\") pod \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.678789 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnlh\" (UniqueName: \"kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh\") pod \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\" (UID: \"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2\") " Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.680700 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities" (OuterVolumeSpecName: "utilities") pod "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" (UID: "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.684964 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh" (OuterVolumeSpecName: "kube-api-access-tpnlh") pod "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" (UID: "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2"). InnerVolumeSpecName "kube-api-access-tpnlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.702719 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" (UID: "d3403a2e-8832-4ed1-a67e-16c6ebd88cd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.779917 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnlh\" (UniqueName: \"kubernetes.io/projected/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-kube-api-access-tpnlh\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.779950 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:04 crc kubenswrapper[4967]: I0120 09:26:04.779960 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.415459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9zfz4" event={"ID":"d3403a2e-8832-4ed1-a67e-16c6ebd88cd2","Type":"ContainerDied","Data":"163b79cc1a3b693258202b9ca664eb663e51d020645bc13a5e7f5dfb3938c2e2"} Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.415515 4967 scope.go:117] "RemoveContainer" containerID="461d4b7cbe26c5345eda98a21997343f527911c3b15a2c50a6bbddc1ca705ec8" Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.415512 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9zfz4" Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.469973 4967 scope.go:117] "RemoveContainer" containerID="322307c020d73e2a592a000c88fa36f175c5664b460775ae1280cd63710b4cb0" Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.475552 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.504275 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9zfz4"] Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.506871 4967 scope.go:117] "RemoveContainer" containerID="8809ab31a7c73679cb9e02aecbb88d495663f0424b693a4b9ec0102a0b84246d" Jan 20 09:26:05 crc kubenswrapper[4967]: I0120 09:26:05.702180 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" path="/var/lib/kubelet/pods/d3403a2e-8832-4ed1-a67e-16c6ebd88cd2/volumes" Jan 20 09:26:08 crc kubenswrapper[4967]: I0120 09:26:08.045210 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:26:08 crc kubenswrapper[4967]: I0120 09:26:08.090154 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.268144 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.440798 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zmlcg" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="registry-server" containerID="cri-o://2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2" gracePeriod=2 Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.790955 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.856779 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities\") pod \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.856900 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content\") pod \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.856934 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n92ft\" (UniqueName: \"kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft\") pod \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\" (UID: \"913ab81d-ce17-4e90-b6cd-7072178ee4ef\") " Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.857823 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities" (OuterVolumeSpecName: "utilities") pod "913ab81d-ce17-4e90-b6cd-7072178ee4ef" (UID: "913ab81d-ce17-4e90-b6cd-7072178ee4ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.862455 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft" (OuterVolumeSpecName: "kube-api-access-n92ft") pod "913ab81d-ce17-4e90-b6cd-7072178ee4ef" (UID: "913ab81d-ce17-4e90-b6cd-7072178ee4ef"). InnerVolumeSpecName "kube-api-access-n92ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.962726 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n92ft\" (UniqueName: \"kubernetes.io/projected/913ab81d-ce17-4e90-b6cd-7072178ee4ef-kube-api-access-n92ft\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.963045 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:09 crc kubenswrapper[4967]: I0120 09:26:09.987428 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "913ab81d-ce17-4e90-b6cd-7072178ee4ef" (UID: "913ab81d-ce17-4e90-b6cd-7072178ee4ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.063965 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913ab81d-ce17-4e90-b6cd-7072178ee4ef-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.449975 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" event={"ID":"561359f4-e525-49a0-b560-d63c5491987b","Type":"ContainerStarted","Data":"b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9"} Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.450034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" event={"ID":"561359f4-e525-49a0-b560-d63c5491987b","Type":"ContainerStarted","Data":"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c"} Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.452563 4967 generic.go:334] "Generic (PLEG): container finished" podID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerID="2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2" exitCode=0 Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.452630 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerDied","Data":"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2"} Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.452673 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zmlcg" event={"ID":"913ab81d-ce17-4e90-b6cd-7072178ee4ef","Type":"ContainerDied","Data":"b24b3b188259c389e2e0b6790884c76409c46b1f69c1a1cdd3e77cd57ce41876"} Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.452701 4967 scope.go:117] "RemoveContainer" containerID="2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.453000 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zmlcg" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.470889 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" podStartSLOduration=1.9153319039999999 podStartE2EDuration="13.470863133s" podCreationTimestamp="2026-01-20 09:25:57 +0000 UTC" firstStartedPulling="2026-01-20 09:25:58.037248785 +0000 UTC m=+2192.622548992" lastFinishedPulling="2026-01-20 09:26:09.592780014 +0000 UTC m=+2204.178080221" observedRunningTime="2026-01-20 09:26:10.46674751 +0000 UTC m=+2205.052047717" watchObservedRunningTime="2026-01-20 09:26:10.470863133 +0000 UTC m=+2205.056163350" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.471942 4967 scope.go:117] "RemoveContainer" containerID="80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.491790 4967 scope.go:117] "RemoveContainer" containerID="9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.493707 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.498112 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zmlcg"] Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.506338 4967 scope.go:117] "RemoveContainer" containerID="2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2" Jan 20 09:26:10 crc kubenswrapper[4967]: E0120 09:26:10.506838 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2\": container with ID starting with 2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2 not found: ID does not exist" containerID="2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.506875 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2"} err="failed to get container status \"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2\": rpc error: code = NotFound desc = could not find container \"2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2\": container with ID starting with 2e4d5be29f79127fdae40f24c6b61c98d2cf3d206d212a1ce5e5c6f6ffdc36e2 not found: ID does not exist" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.506901 4967 scope.go:117] "RemoveContainer" containerID="80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516" Jan 20 09:26:10 crc kubenswrapper[4967]: E0120 09:26:10.507260 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516\": container with ID starting with 80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516 not found: ID does not exist" containerID="80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.507336 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516"} err="failed to get container status \"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516\": rpc error: code = NotFound desc = could not find container \"80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516\": container with ID starting with 80ac23089a76de8dd84d5989a9da52e94858f31072e4d705591ce224cf05d516 not found: ID does not exist" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.507399 4967 scope.go:117] "RemoveContainer" containerID="9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db" Jan 20 09:26:10 crc kubenswrapper[4967]: E0120 09:26:10.507707 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db\": container with ID starting with 9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db not found: ID does not exist" containerID="9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db" Jan 20 09:26:10 crc kubenswrapper[4967]: I0120 09:26:10.507798 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db"} err="failed to get container status \"9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db\": rpc error: code = NotFound desc = could not find container \"9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db\": container with ID starting with 9b3d9b0f04fc79a8581e2781da707a26b0b7afb763942c4a4b04bc7841c6d2db not found: ID does not exist" Jan 20 09:26:11 crc kubenswrapper[4967]: I0120 09:26:11.702482 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" path="/var/lib/kubelet/pods/913ab81d-ce17-4e90-b6cd-7072178ee4ef/volumes" Jan 20 09:26:18 crc kubenswrapper[4967]: I0120 09:26:18.474462 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:26:18 crc kubenswrapper[4967]: I0120 09:26:18.474855 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.640267 4967 scope.go:117] "RemoveContainer" containerID="99254b5141c34acb9f76f0723e298d765c9033b7f9ba283efa5b225b46291a7f" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.670138 4967 scope.go:117] "RemoveContainer" containerID="bf758abfee8d25dd67d8d7924ed8eb925e4aa8e06c57021a1fe24a3aff461a48" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.690120 4967 scope.go:117] "RemoveContainer" containerID="793c635cc9edb7856b11ecec42c20a318a1c5ccb3167f75e629112d9bc413229" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.712307 4967 scope.go:117] "RemoveContainer" containerID="f4b8eb2d6e37d1673728a01f88ac6deea3102110aad10b7a469da0c015bc751f" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.734798 4967 scope.go:117] "RemoveContainer" containerID="1d0e51ee119e72be2320cbac89d3ddc1d45ee1d4d73e1caf6c7380f209db596e" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.753630 4967 scope.go:117] "RemoveContainer" containerID="839283610dc65f6081a084c2ecf836fc16beb1b7f9a03d6eeb8a0875e5196b8d" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.774076 4967 scope.go:117] "RemoveContainer" containerID="69ded61430a547c32b4a3941a4ce7861ee7051fe4ad365a827603cfe78e49077" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.788424 4967 scope.go:117] "RemoveContainer" containerID="7bd938a5ab03de0a27d4c58531d093380d90cdfa7b4aed9834e1c842c06b7009" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.804026 4967 scope.go:117] "RemoveContainer" containerID="2c4816b994da454fa5dabde287637959a4064f487a65bcc70c63892942a32dd6" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.825902 4967 scope.go:117] "RemoveContainer" containerID="8781c51400de670a71aa7c6699e57908566242055de233b4237a6e9bafb676b4" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.843159 4967 scope.go:117] "RemoveContainer" containerID="118445c68a573eb23fe42c671e11f042b7d6f627aa12473582d6379eb23eadc1" Jan 20 09:26:27 crc kubenswrapper[4967]: I0120 09:26:27.871479 4967 scope.go:117] "RemoveContainer" containerID="e47033e9c8290a2c0e606c58e8a59d063f0252cad126658b578380d065358ce4" Jan 20 09:26:48 crc kubenswrapper[4967]: I0120 09:26:48.474990 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:26:48 crc kubenswrapper[4967]: I0120 09:26:48.475588 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:26:50 crc kubenswrapper[4967]: I0120 09:26:50.431545 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5gqsg_cc8e9166-fdeb-4308-a44e-e932397bfead/control-plane-machine-set-operator/0.log" Jan 20 09:26:50 crc kubenswrapper[4967]: I0120 09:26:50.597390 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gpj5p_f6f75790-6be7-4f4c-ae17-5258a715666e/machine-api-operator/0.log" Jan 20 09:26:50 crc kubenswrapper[4967]: I0120 09:26:50.623293 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gpj5p_f6f75790-6be7-4f4c-ae17-5258a715666e/kube-rbac-proxy/0.log" Jan 20 09:27:17 crc kubenswrapper[4967]: I0120 09:27:17.878032 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-gzh9x_180c30fb-af36-47ef-87dc-8d635f6a9e4d/kube-rbac-proxy/0.log" Jan 20 09:27:17 crc kubenswrapper[4967]: I0120 09:27:17.908156 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-gzh9x_180c30fb-af36-47ef-87dc-8d635f6a9e4d/controller/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.030093 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.206675 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.225799 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.232637 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.264118 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.420789 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.420838 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.434313 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.457122 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.474685 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.474743 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.474787 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.475309 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.475362 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" gracePeriod=600 Jan 20 09:27:18 crc kubenswrapper[4967]: E0120 09:27:18.600491 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.611695 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.623201 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.632724 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.658299 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/controller/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.822842 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/kube-rbac-proxy/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.839453 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/frr-metrics/0.log" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.851415 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" exitCode=0 Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.851454 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723"} Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.851483 4967 scope.go:117] "RemoveContainer" containerID="9f99817c5f153c3f1d8c55bb4507567f5389329e38191cc10458fdb6b8ffa5dc" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.852456 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:27:18 crc kubenswrapper[4967]: E0120 09:27:18.863848 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:27:18 crc kubenswrapper[4967]: I0120 09:27:18.881982 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/kube-rbac-proxy-frr/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.000743 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/reloader/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.058101 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-q72g4_87e441f9-d588-4f98-ad15-d632b239aef5/frr-k8s-webhook-server/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.196416 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/frr/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.220313 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9c99dbd64-q2zrc_ccc3a123-c88b-42ae-8c10-5fc70762b0a3/manager/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.337194 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74dcbc6877-8v4vv_96b5beda-b9ed-4f48-9dcb-8a261f532b47/webhook-server/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.381938 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hqtb9_0ac95690-ca20-424c-af78-4134a67a01a2/kube-rbac-proxy/0.log" Jan 20 09:27:19 crc kubenswrapper[4967]: I0120 09:27:19.533661 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hqtb9_0ac95690-ca20-424c-af78-4134a67a01a2/speaker/0.log" Jan 20 09:27:30 crc kubenswrapper[4967]: I0120 09:27:30.694579 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:27:30 crc kubenswrapper[4967]: E0120 09:27:30.695156 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.011738 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.166402 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.189755 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.235092 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.351737 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.369793 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.382763 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/extract/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.509326 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-utilities/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.649521 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-utilities/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.657669 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-content/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.661792 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-content/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.694024 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:27:42 crc kubenswrapper[4967]: E0120 09:27:42.694380 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.817680 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-utilities/0.log" Jan 20 09:27:42 crc kubenswrapper[4967]: I0120 09:27:42.820823 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.035088 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.169545 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-s5trl_4baaa2ac-2a77-4886-9417-da415f5119ff/registry-server/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.210125 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.233467 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.236503 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.348845 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.382200 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.570331 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.598228 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sqc5w_778bd2b6-c30e-43eb-8045-ff0720482f5c/marketplace-operator/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.832507 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.861496 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.878967 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:27:43 crc kubenswrapper[4967]: I0120 09:27:43.962692 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/registry-server/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.045809 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.107096 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.148564 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/registry-server/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.252146 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.399534 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.420587 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.434913 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.589319 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:27:44 crc kubenswrapper[4967]: I0120 09:27:44.594446 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:27:45 crc kubenswrapper[4967]: I0120 09:27:45.044214 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/registry-server/0.log" Jan 20 09:27:54 crc kubenswrapper[4967]: I0120 09:27:54.694793 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:27:54 crc kubenswrapper[4967]: E0120 09:27:54.695636 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:28:08 crc kubenswrapper[4967]: I0120 09:28:08.694751 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:28:08 crc kubenswrapper[4967]: E0120 09:28:08.696376 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:28:19 crc kubenswrapper[4967]: I0120 09:28:19.694312 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:28:19 crc kubenswrapper[4967]: E0120 09:28:19.695015 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:28:34 crc kubenswrapper[4967]: I0120 09:28:34.694832 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:28:34 crc kubenswrapper[4967]: E0120 09:28:34.695716 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:28:49 crc kubenswrapper[4967]: I0120 09:28:49.694329 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:28:49 crc kubenswrapper[4967]: E0120 09:28:49.696720 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:28:56 crc kubenswrapper[4967]: I0120 09:28:56.436746 4967 generic.go:334] "Generic (PLEG): container finished" podID="561359f4-e525-49a0-b560-d63c5491987b" containerID="ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c" exitCode=0 Jan 20 09:28:56 crc kubenswrapper[4967]: I0120 09:28:56.436826 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" event={"ID":"561359f4-e525-49a0-b560-d63c5491987b","Type":"ContainerDied","Data":"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c"} Jan 20 09:28:56 crc kubenswrapper[4967]: I0120 09:28:56.437658 4967 scope.go:117] "RemoveContainer" containerID="ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c" Jan 20 09:28:57 crc kubenswrapper[4967]: I0120 09:28:57.363276 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wp7vg_must-gather-r4fvk_561359f4-e525-49a0-b560-d63c5491987b/gather/0.log" Jan 20 09:29:01 crc kubenswrapper[4967]: I0120 09:29:01.693887 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:29:01 crc kubenswrapper[4967]: E0120 09:29:01.694796 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:29:04 crc kubenswrapper[4967]: I0120 09:29:04.635574 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wp7vg/must-gather-r4fvk"] Jan 20 09:29:04 crc kubenswrapper[4967]: I0120 09:29:04.637353 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="copy" containerID="cri-o://b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9" gracePeriod=2 Jan 20 09:29:04 crc kubenswrapper[4967]: I0120 09:29:04.645281 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wp7vg/must-gather-r4fvk"] Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.477552 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wp7vg_must-gather-r4fvk_561359f4-e525-49a0-b560-d63c5491987b/copy/0.log" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.478663 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.496949 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wp7vg_must-gather-r4fvk_561359f4-e525-49a0-b560-d63c5491987b/copy/0.log" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.497553 4967 generic.go:334] "Generic (PLEG): container finished" podID="561359f4-e525-49a0-b560-d63c5491987b" containerID="b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9" exitCode=143 Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.497634 4967 scope.go:117] "RemoveContainer" containerID="b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.497649 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wp7vg/must-gather-r4fvk" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.515390 4967 scope.go:117] "RemoveContainer" containerID="ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.541642 4967 scope.go:117] "RemoveContainer" containerID="b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9" Jan 20 09:29:05 crc kubenswrapper[4967]: E0120 09:29:05.542170 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9\": container with ID starting with b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9 not found: ID does not exist" containerID="b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.542239 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9"} err="failed to get container status \"b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9\": rpc error: code = NotFound desc = could not find container \"b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9\": container with ID starting with b536b1bfd2141a248db32d289576acd2d4250cf16d1268acbff08e4060f0c7d9 not found: ID does not exist" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.542286 4967 scope.go:117] "RemoveContainer" containerID="ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c" Jan 20 09:29:05 crc kubenswrapper[4967]: E0120 09:29:05.542858 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c\": container with ID starting with ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c not found: ID does not exist" containerID="ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.542918 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c"} err="failed to get container status \"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c\": rpc error: code = NotFound desc = could not find container \"ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c\": container with ID starting with ab3f014673d5701d781206df6e2375ef58076e5845120fb9dbc5ab789ad3e00c not found: ID does not exist" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.643871 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z78r4\" (UniqueName: \"kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4\") pod \"561359f4-e525-49a0-b560-d63c5491987b\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.644007 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output\") pod \"561359f4-e525-49a0-b560-d63c5491987b\" (UID: \"561359f4-e525-49a0-b560-d63c5491987b\") " Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.662447 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4" (OuterVolumeSpecName: "kube-api-access-z78r4") pod "561359f4-e525-49a0-b560-d63c5491987b" (UID: "561359f4-e525-49a0-b560-d63c5491987b"). InnerVolumeSpecName "kube-api-access-z78r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.718833 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "561359f4-e525-49a0-b560-d63c5491987b" (UID: "561359f4-e525-49a0-b560-d63c5491987b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.747893 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z78r4\" (UniqueName: \"kubernetes.io/projected/561359f4-e525-49a0-b560-d63c5491987b-kube-api-access-z78r4\") on node \"crc\" DevicePath \"\"" Jan 20 09:29:05 crc kubenswrapper[4967]: I0120 09:29:05.747937 4967 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/561359f4-e525-49a0-b560-d63c5491987b-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 20 09:29:07 crc kubenswrapper[4967]: I0120 09:29:07.701194 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561359f4-e525-49a0-b560-d63c5491987b" path="/var/lib/kubelet/pods/561359f4-e525-49a0-b560-d63c5491987b/volumes" Jan 20 09:29:14 crc kubenswrapper[4967]: I0120 09:29:14.694765 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:29:14 crc kubenswrapper[4967]: E0120 09:29:14.695633 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:29:26 crc kubenswrapper[4967]: I0120 09:29:26.695076 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:29:26 crc kubenswrapper[4967]: E0120 09:29:26.696078 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:29:38 crc kubenswrapper[4967]: I0120 09:29:38.694575 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:29:38 crc kubenswrapper[4967]: E0120 09:29:38.695396 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:29:50 crc kubenswrapper[4967]: I0120 09:29:50.694475 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:29:50 crc kubenswrapper[4967]: E0120 09:29:50.695353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.152680 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2"] Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153278 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153295 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153308 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="extract-utilities" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153317 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="extract-utilities" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153338 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="copy" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153346 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="copy" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153356 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153363 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153372 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="extract-content" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153379 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="extract-content" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153389 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="extract-content" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153396 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="extract-content" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153406 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="extract-utilities" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153412 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="extract-utilities" Jan 20 09:30:00 crc kubenswrapper[4967]: E0120 09:30:00.153420 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="gather" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153426 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="gather" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153558 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3403a2e-8832-4ed1-a67e-16c6ebd88cd2" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153572 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="913ab81d-ce17-4e90-b6cd-7072178ee4ef" containerName="registry-server" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153582 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="gather" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.153596 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="561359f4-e525-49a0-b560-d63c5491987b" containerName="copy" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.154143 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.156960 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.157083 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.169939 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2"] Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.201540 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.201900 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jxq\" (UniqueName: \"kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.201963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.303586 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.303721 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jxq\" (UniqueName: \"kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.303755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.304653 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.309594 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.319099 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jxq\" (UniqueName: \"kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq\") pod \"collect-profiles-29481690-q6ff2\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.476891 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:00 crc kubenswrapper[4967]: I0120 09:30:00.904498 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2"] Jan 20 09:30:01 crc kubenswrapper[4967]: I0120 09:30:01.841199 4967 generic.go:334] "Generic (PLEG): container finished" podID="d955b7f4-5b5d-4e54-8adb-5de789fc02bf" containerID="0a64f0529495a105df64ace3907ec1e1619690af241e73392ed7cdf6524bec51" exitCode=0 Jan 20 09:30:01 crc kubenswrapper[4967]: I0120 09:30:01.841287 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" event={"ID":"d955b7f4-5b5d-4e54-8adb-5de789fc02bf","Type":"ContainerDied","Data":"0a64f0529495a105df64ace3907ec1e1619690af241e73392ed7cdf6524bec51"} Jan 20 09:30:01 crc kubenswrapper[4967]: I0120 09:30:01.841471 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" event={"ID":"d955b7f4-5b5d-4e54-8adb-5de789fc02bf","Type":"ContainerStarted","Data":"a73fc3085f82925b1fe3a6a88ff1f0c605ff594cfa283cbc80c56241ee5e8a43"} Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.111241 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.243322 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94jxq\" (UniqueName: \"kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq\") pod \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.243469 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume\") pod \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.243509 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume\") pod \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\" (UID: \"d955b7f4-5b5d-4e54-8adb-5de789fc02bf\") " Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.244318 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume" (OuterVolumeSpecName: "config-volume") pod "d955b7f4-5b5d-4e54-8adb-5de789fc02bf" (UID: "d955b7f4-5b5d-4e54-8adb-5de789fc02bf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.248834 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq" (OuterVolumeSpecName: "kube-api-access-94jxq") pod "d955b7f4-5b5d-4e54-8adb-5de789fc02bf" (UID: "d955b7f4-5b5d-4e54-8adb-5de789fc02bf"). InnerVolumeSpecName "kube-api-access-94jxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.249709 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d955b7f4-5b5d-4e54-8adb-5de789fc02bf" (UID: "d955b7f4-5b5d-4e54-8adb-5de789fc02bf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.344952 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-config-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.345002 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.345017 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94jxq\" (UniqueName: \"kubernetes.io/projected/d955b7f4-5b5d-4e54-8adb-5de789fc02bf-kube-api-access-94jxq\") on node \"crc\" DevicePath \"\"" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.865828 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" event={"ID":"d955b7f4-5b5d-4e54-8adb-5de789fc02bf","Type":"ContainerDied","Data":"a73fc3085f82925b1fe3a6a88ff1f0c605ff594cfa283cbc80c56241ee5e8a43"} Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.866120 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73fc3085f82925b1fe3a6a88ff1f0c605ff594cfa283cbc80c56241ee5e8a43" Jan 20 09:30:03 crc kubenswrapper[4967]: I0120 09:30:03.866280 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29481690-q6ff2" Jan 20 09:30:04 crc kubenswrapper[4967]: I0120 09:30:04.186670 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v"] Jan 20 09:30:04 crc kubenswrapper[4967]: I0120 09:30:04.190857 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29481645-m6h5v"] Jan 20 09:30:05 crc kubenswrapper[4967]: I0120 09:30:05.696893 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:30:05 crc kubenswrapper[4967]: E0120 09:30:05.697115 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:30:05 crc kubenswrapper[4967]: I0120 09:30:05.703521 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="334064e9-09d2-4cd7-8eb4-75449c214712" path="/var/lib/kubelet/pods/334064e9-09d2-4cd7-8eb4-75449c214712/volumes" Jan 20 09:30:19 crc kubenswrapper[4967]: I0120 09:30:19.694546 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:30:19 crc kubenswrapper[4967]: E0120 09:30:19.695191 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:30:28 crc kubenswrapper[4967]: I0120 09:30:28.177288 4967 scope.go:117] "RemoveContainer" containerID="73533bd151607d051da43e6dacce101a24f0fd41e3a5417c5b73a2f2fc0cbe11" Jan 20 09:30:31 crc kubenswrapper[4967]: I0120 09:30:31.693921 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:30:31 crc kubenswrapper[4967]: E0120 09:30:31.694411 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:30:44 crc kubenswrapper[4967]: I0120 09:30:44.694423 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:30:44 crc kubenswrapper[4967]: E0120 09:30:44.695087 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:30:56 crc kubenswrapper[4967]: I0120 09:30:56.694797 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:30:56 crc kubenswrapper[4967]: E0120 09:30:56.696112 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:31:08 crc kubenswrapper[4967]: I0120 09:31:08.694394 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:31:08 crc kubenswrapper[4967]: E0120 09:31:08.695115 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:31:21 crc kubenswrapper[4967]: I0120 09:31:21.693967 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:31:21 crc kubenswrapper[4967]: E0120 09:31:21.694794 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:31:28 crc kubenswrapper[4967]: I0120 09:31:28.209283 4967 scope.go:117] "RemoveContainer" containerID="c8333d326331a494cdd703a528b99b8a1c57448d6654d981371357ccda60a8e6" Jan 20 09:31:34 crc kubenswrapper[4967]: I0120 09:31:34.693985 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:31:34 crc kubenswrapper[4967]: E0120 09:31:34.694579 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:31:47 crc kubenswrapper[4967]: I0120 09:31:47.694538 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:31:47 crc kubenswrapper[4967]: E0120 09:31:47.695251 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.007236 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4ht7m/must-gather-z6djj"] Jan 20 09:31:55 crc kubenswrapper[4967]: E0120 09:31:55.008073 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d955b7f4-5b5d-4e54-8adb-5de789fc02bf" containerName="collect-profiles" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.008093 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d955b7f4-5b5d-4e54-8adb-5de789fc02bf" containerName="collect-profiles" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.008232 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d955b7f4-5b5d-4e54-8adb-5de789fc02bf" containerName="collect-profiles" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.008988 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.014892 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4ht7m"/"kube-root-ca.crt" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.017191 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4ht7m"/"openshift-service-ca.crt" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.026817 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4ht7m/must-gather-z6djj"] Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.110836 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.110881 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhpv\" (UniqueName: \"kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.212156 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.212478 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhpv\" (UniqueName: \"kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.212648 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.235328 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhpv\" (UniqueName: \"kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv\") pod \"must-gather-z6djj\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.328699 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:31:55 crc kubenswrapper[4967]: I0120 09:31:55.737131 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4ht7m/must-gather-z6djj"] Jan 20 09:31:56 crc kubenswrapper[4967]: I0120 09:31:56.532553 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4ht7m/must-gather-z6djj" event={"ID":"b86d9182-cb86-4c2d-87b0-c0e537e100e5","Type":"ContainerStarted","Data":"22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16"} Jan 20 09:31:56 crc kubenswrapper[4967]: I0120 09:31:56.532874 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4ht7m/must-gather-z6djj" event={"ID":"b86d9182-cb86-4c2d-87b0-c0e537e100e5","Type":"ContainerStarted","Data":"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05"} Jan 20 09:31:56 crc kubenswrapper[4967]: I0120 09:31:56.532885 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4ht7m/must-gather-z6djj" event={"ID":"b86d9182-cb86-4c2d-87b0-c0e537e100e5","Type":"ContainerStarted","Data":"f6c75e2820a5d00c14497b152204e264420fd6c22dc281530e70927d75ee2ef4"} Jan 20 09:31:56 crc kubenswrapper[4967]: I0120 09:31:56.547776 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4ht7m/must-gather-z6djj" podStartSLOduration=2.5477593020000002 podStartE2EDuration="2.547759302s" podCreationTimestamp="2026-01-20 09:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-20 09:31:56.545754727 +0000 UTC m=+2551.131054944" watchObservedRunningTime="2026-01-20 09:31:56.547759302 +0000 UTC m=+2551.133059509" Jan 20 09:32:02 crc kubenswrapper[4967]: I0120 09:32:02.694519 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:32:02 crc kubenswrapper[4967]: E0120 09:32:02.695261 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:32:16 crc kubenswrapper[4967]: I0120 09:32:16.694538 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:32:16 crc kubenswrapper[4967]: E0120 09:32:16.696176 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:32:29 crc kubenswrapper[4967]: I0120 09:32:29.694486 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.456555 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7p2cf"] Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.458504 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.473811 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p2cf"] Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.502449 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-catalog-content\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.502551 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-utilities\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.502579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vppj7\" (UniqueName: \"kubernetes.io/projected/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-kube-api-access-vppj7\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.603451 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-catalog-content\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.603522 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-utilities\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.603544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vppj7\" (UniqueName: \"kubernetes.io/projected/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-kube-api-access-vppj7\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.604018 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-catalog-content\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.604117 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-utilities\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.641135 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vppj7\" (UniqueName: \"kubernetes.io/projected/acd49685-4627-44ac-a5ce-3e7e4ba4f0a8-kube-api-access-vppj7\") pod \"certified-operators-7p2cf\" (UID: \"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8\") " pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.749799 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c"} Jan 20 09:32:30 crc kubenswrapper[4967]: I0120 09:32:30.837159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:31 crc kubenswrapper[4967]: W0120 09:32:31.298784 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacd49685_4627_44ac_a5ce_3e7e4ba4f0a8.slice/crio-7b4e04df12c588d3270f6bc369d48a36f642f763f7cd7d863eff95a2017402b4 WatchSource:0}: Error finding container 7b4e04df12c588d3270f6bc369d48a36f642f763f7cd7d863eff95a2017402b4: Status 404 returned error can't find the container with id 7b4e04df12c588d3270f6bc369d48a36f642f763f7cd7d863eff95a2017402b4 Jan 20 09:32:31 crc kubenswrapper[4967]: I0120 09:32:31.314244 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p2cf"] Jan 20 09:32:31 crc kubenswrapper[4967]: I0120 09:32:31.756352 4967 generic.go:334] "Generic (PLEG): container finished" podID="acd49685-4627-44ac-a5ce-3e7e4ba4f0a8" containerID="3c9c392624677b1982b0e55bed8d289dd428c6a296b3b607f29cfa7518d2c039" exitCode=0 Jan 20 09:32:31 crc kubenswrapper[4967]: I0120 09:32:31.756905 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2cf" event={"ID":"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8","Type":"ContainerDied","Data":"3c9c392624677b1982b0e55bed8d289dd428c6a296b3b607f29cfa7518d2c039"} Jan 20 09:32:31 crc kubenswrapper[4967]: I0120 09:32:31.756937 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2cf" event={"ID":"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8","Type":"ContainerStarted","Data":"7b4e04df12c588d3270f6bc369d48a36f642f763f7cd7d863eff95a2017402b4"} Jan 20 09:32:31 crc kubenswrapper[4967]: I0120 09:32:31.758254 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 20 09:32:37 crc kubenswrapper[4967]: I0120 09:32:37.004461 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5gqsg_cc8e9166-fdeb-4308-a44e-e932397bfead/control-plane-machine-set-operator/0.log" Jan 20 09:32:37 crc kubenswrapper[4967]: I0120 09:32:37.158796 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gpj5p_f6f75790-6be7-4f4c-ae17-5258a715666e/kube-rbac-proxy/0.log" Jan 20 09:32:37 crc kubenswrapper[4967]: I0120 09:32:37.205015 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gpj5p_f6f75790-6be7-4f4c-ae17-5258a715666e/machine-api-operator/0.log" Jan 20 09:32:37 crc kubenswrapper[4967]: I0120 09:32:37.803177 4967 generic.go:334] "Generic (PLEG): container finished" podID="acd49685-4627-44ac-a5ce-3e7e4ba4f0a8" containerID="7b09fd2959d12c89417232bc00bcb073f2119d82d8b7941c190b93dd11229090" exitCode=0 Jan 20 09:32:37 crc kubenswrapper[4967]: I0120 09:32:37.803224 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2cf" event={"ID":"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8","Type":"ContainerDied","Data":"7b09fd2959d12c89417232bc00bcb073f2119d82d8b7941c190b93dd11229090"} Jan 20 09:32:39 crc kubenswrapper[4967]: I0120 09:32:39.818293 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2cf" event={"ID":"acd49685-4627-44ac-a5ce-3e7e4ba4f0a8","Type":"ContainerStarted","Data":"85aaf22701ccf2340499dce2f5c3f86e607996fe6efd84dc95e8cb655ea5dc8f"} Jan 20 09:32:39 crc kubenswrapper[4967]: I0120 09:32:39.842638 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7p2cf" podStartSLOduration=2.826461289 podStartE2EDuration="9.842601587s" podCreationTimestamp="2026-01-20 09:32:30 +0000 UTC" firstStartedPulling="2026-01-20 09:32:31.757970797 +0000 UTC m=+2586.343271004" lastFinishedPulling="2026-01-20 09:32:38.774111085 +0000 UTC m=+2593.359411302" observedRunningTime="2026-01-20 09:32:39.839843461 +0000 UTC m=+2594.425143668" watchObservedRunningTime="2026-01-20 09:32:39.842601587 +0000 UTC m=+2594.427901794" Jan 20 09:32:40 crc kubenswrapper[4967]: I0120 09:32:40.839037 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:40 crc kubenswrapper[4967]: I0120 09:32:40.839104 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:40 crc kubenswrapper[4967]: I0120 09:32:40.881209 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:50 crc kubenswrapper[4967]: I0120 09:32:50.879578 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7p2cf" Jan 20 09:32:50 crc kubenswrapper[4967]: I0120 09:32:50.954903 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p2cf"] Jan 20 09:32:50 crc kubenswrapper[4967]: I0120 09:32:50.988885 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 09:32:50 crc kubenswrapper[4967]: I0120 09:32:50.989216 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s5trl" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="registry-server" containerID="cri-o://8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2" gracePeriod=2 Jan 20 09:32:51 crc kubenswrapper[4967]: E0120 09:32:51.123015 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4baaa2ac_2a77_4886_9417_da415f5119ff.slice/crio-8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4baaa2ac_2a77_4886_9417_da415f5119ff.slice/crio-conmon-8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2.scope\": RecentStats: unable to find data in memory cache]" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.357464 4967 generic.go:334] "Generic (PLEG): container finished" podID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerID="8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2" exitCode=0 Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.358492 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerDied","Data":"8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2"} Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.358520 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5trl" event={"ID":"4baaa2ac-2a77-4886-9417-da415f5119ff","Type":"ContainerDied","Data":"d614a956ee9c2f9f4a8c6e74a0731c8297c629f48081a445b0ba5112f3a50f94"} Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.358530 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d614a956ee9c2f9f4a8c6e74a0731c8297c629f48081a445b0ba5112f3a50f94" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.372767 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.472283 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content\") pod \"4baaa2ac-2a77-4886-9417-da415f5119ff\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.472424 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities\") pod \"4baaa2ac-2a77-4886-9417-da415f5119ff\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.472447 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkrpr\" (UniqueName: \"kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr\") pod \"4baaa2ac-2a77-4886-9417-da415f5119ff\" (UID: \"4baaa2ac-2a77-4886-9417-da415f5119ff\") " Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.473180 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities" (OuterVolumeSpecName: "utilities") pod "4baaa2ac-2a77-4886-9417-da415f5119ff" (UID: "4baaa2ac-2a77-4886-9417-da415f5119ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.477182 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr" (OuterVolumeSpecName: "kube-api-access-zkrpr") pod "4baaa2ac-2a77-4886-9417-da415f5119ff" (UID: "4baaa2ac-2a77-4886-9417-da415f5119ff"). InnerVolumeSpecName "kube-api-access-zkrpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.526157 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4baaa2ac-2a77-4886-9417-da415f5119ff" (UID: "4baaa2ac-2a77-4886-9417-da415f5119ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.574190 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.574226 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkrpr\" (UniqueName: \"kubernetes.io/projected/4baaa2ac-2a77-4886-9417-da415f5119ff-kube-api-access-zkrpr\") on node \"crc\" DevicePath \"\"" Jan 20 09:32:51 crc kubenswrapper[4967]: I0120 09:32:51.574257 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4baaa2ac-2a77-4886-9417-da415f5119ff-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:32:52 crc kubenswrapper[4967]: I0120 09:32:52.361934 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5trl" Jan 20 09:32:52 crc kubenswrapper[4967]: I0120 09:32:52.381750 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 09:32:52 crc kubenswrapper[4967]: I0120 09:32:52.387274 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s5trl"] Jan 20 09:32:53 crc kubenswrapper[4967]: I0120 09:32:53.702962 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" path="/var/lib/kubelet/pods/4baaa2ac-2a77-4886-9417-da415f5119ff/volumes" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.084726 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-gzh9x_180c30fb-af36-47ef-87dc-8d635f6a9e4d/controller/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.110010 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-gzh9x_180c30fb-af36-47ef-87dc-8d635f6a9e4d/kube-rbac-proxy/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.399563 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.570705 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.603529 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.603718 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.629794 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.766671 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.787563 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.806098 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.820288 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:33:05 crc kubenswrapper[4967]: I0120 09:33:05.988632 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-frr-files/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.013370 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-metrics/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.020181 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/cp-reloader/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.040899 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/controller/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.172703 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/kube-rbac-proxy/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.217061 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/kube-rbac-proxy-frr/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.233682 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/frr-metrics/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.343131 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/reloader/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.495679 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-q72g4_87e441f9-d588-4f98-ad15-d632b239aef5/frr-k8s-webhook-server/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.610002 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9c99dbd64-q2zrc_ccc3a123-c88b-42ae-8c10-5fc70762b0a3/manager/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.681627 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74dcbc6877-8v4vv_96b5beda-b9ed-4f48-9dcb-8a261f532b47/webhook-server/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.690032 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96vmf_0deabec7-6e63-4152-9107-1045d11c8e6c/frr/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.883769 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hqtb9_0ac95690-ca20-424c-af78-4134a67a01a2/kube-rbac-proxy/0.log" Jan 20 09:33:06 crc kubenswrapper[4967]: I0120 09:33:06.912501 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hqtb9_0ac95690-ca20-424c-af78-4134a67a01a2/speaker/0.log" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.275022 4967 scope.go:117] "RemoveContainer" containerID="8d740f7d0d9cd5b9a22b43c90cbedeaf292b7560e4d36d3b4ea1210f9f6511b2" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.294805 4967 scope.go:117] "RemoveContainer" containerID="86b8c57e65bbd3470c687010030f2558bdd380dfc15272e8859d48ace8f88a95" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.308964 4967 scope.go:117] "RemoveContainer" containerID="160d3324cfae8fbba6d67a20085ad7ac24d9cf4944d6bf411a0d5ea7dfddc188" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.840931 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.959548 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:33:28 crc kubenswrapper[4967]: I0120 09:33:28.994805 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.002122 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.202020 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/pull/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.202087 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/extract/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.234718 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dccvv5c_5de711c7-4972-4b89-8842-8d70a4126c13/util/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.360252 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-utilities/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.507936 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-utilities/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.583317 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-content/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.600257 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-content/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.721635 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-content/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.732058 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/extract-utilities/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.883894 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7p2cf_acd49685-4627-44ac-a5ce-3e7e4ba4f0a8/registry-server/0.log" Jan 20 09:33:29 crc kubenswrapper[4967]: I0120 09:33:29.919767 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.068438 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.098787 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.100553 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.360746 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-utilities/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.400682 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/extract-content/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.562012 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sqc5w_778bd2b6-c30e-43eb-8045-ff0720482f5c/marketplace-operator/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.812690 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.980082 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.990820 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xhnv8_a2774890-f4e7-47b2-97c2-d308e78b7256/registry-server/0.log" Jan 20 09:33:30 crc kubenswrapper[4967]: I0120 09:33:30.999044 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.022867 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.165965 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-utilities/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.188356 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/extract-content/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.260594 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f9v7m_76857b90-e875-42ff-9332-a775e9cc5af6/registry-server/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.370640 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.555596 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.559713 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.567135 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.760080 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-utilities/0.log" Jan 20 09:33:31 crc kubenswrapper[4967]: I0120 09:33:31.810197 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/extract-content/0.log" Jan 20 09:33:32 crc kubenswrapper[4967]: I0120 09:33:32.255446 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q6d4w_b1cfca35-a47d-4fc4-a425-57a60dd4b855/registry-server/0.log" Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.022988 4967 generic.go:334] "Generic (PLEG): container finished" podID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerID="c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05" exitCode=0 Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.023078 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4ht7m/must-gather-z6djj" event={"ID":"b86d9182-cb86-4c2d-87b0-c0e537e100e5","Type":"ContainerDied","Data":"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05"} Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.024033 4967 scope.go:117] "RemoveContainer" containerID="c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05" Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.197538 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4ht7m_must-gather-z6djj_b86d9182-cb86-4c2d-87b0-c0e537e100e5/gather/0.log" Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.474766 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:34:48 crc kubenswrapper[4967]: I0120 09:34:48.474830 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.184766 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4ht7m/must-gather-z6djj"] Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.185539 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4ht7m/must-gather-z6djj" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="copy" containerID="cri-o://22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16" gracePeriod=2 Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.191340 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4ht7m/must-gather-z6djj"] Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.509752 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4ht7m_must-gather-z6djj_b86d9182-cb86-4c2d-87b0-c0e537e100e5/copy/0.log" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.510247 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.591551 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output\") pod \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.591730 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfhpv\" (UniqueName: \"kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv\") pod \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\" (UID: \"b86d9182-cb86-4c2d-87b0-c0e537e100e5\") " Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.598059 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv" (OuterVolumeSpecName: "kube-api-access-tfhpv") pod "b86d9182-cb86-4c2d-87b0-c0e537e100e5" (UID: "b86d9182-cb86-4c2d-87b0-c0e537e100e5"). InnerVolumeSpecName "kube-api-access-tfhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.664279 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b86d9182-cb86-4c2d-87b0-c0e537e100e5" (UID: "b86d9182-cb86-4c2d-87b0-c0e537e100e5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.693075 4967 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b86d9182-cb86-4c2d-87b0-c0e537e100e5-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.693117 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfhpv\" (UniqueName: \"kubernetes.io/projected/b86d9182-cb86-4c2d-87b0-c0e537e100e5-kube-api-access-tfhpv\") on node \"crc\" DevicePath \"\"" Jan 20 09:34:57 crc kubenswrapper[4967]: I0120 09:34:57.702063 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" path="/var/lib/kubelet/pods/b86d9182-cb86-4c2d-87b0-c0e537e100e5/volumes" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.085789 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4ht7m_must-gather-z6djj_b86d9182-cb86-4c2d-87b0-c0e537e100e5/copy/0.log" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.086342 4967 generic.go:334] "Generic (PLEG): container finished" podID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerID="22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16" exitCode=143 Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.086394 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4ht7m/must-gather-z6djj" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.086405 4967 scope.go:117] "RemoveContainer" containerID="22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.104079 4967 scope.go:117] "RemoveContainer" containerID="c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.132332 4967 scope.go:117] "RemoveContainer" containerID="22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16" Jan 20 09:34:58 crc kubenswrapper[4967]: E0120 09:34:58.132889 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16\": container with ID starting with 22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16 not found: ID does not exist" containerID="22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.132931 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16"} err="failed to get container status \"22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16\": rpc error: code = NotFound desc = could not find container \"22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16\": container with ID starting with 22b7b8b4a49ae9ef79d290b1818c9ce5a28a8d8408104a267e9a7f0e4dde5d16 not found: ID does not exist" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.132957 4967 scope.go:117] "RemoveContainer" containerID="c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05" Jan 20 09:34:58 crc kubenswrapper[4967]: E0120 09:34:58.133249 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05\": container with ID starting with c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05 not found: ID does not exist" containerID="c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05" Jan 20 09:34:58 crc kubenswrapper[4967]: I0120 09:34:58.133281 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05"} err="failed to get container status \"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05\": rpc error: code = NotFound desc = could not find container \"c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05\": container with ID starting with c634daa83c98bdc1a943660b07eb5d118ad49b3d96d5b60c65c97ae6bec7fd05 not found: ID does not exist" Jan 20 09:35:18 crc kubenswrapper[4967]: I0120 09:35:18.474017 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:35:18 crc kubenswrapper[4967]: I0120 09:35:18.474568 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.830533 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:39 crc kubenswrapper[4967]: E0120 09:35:39.831660 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="extract-content" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831683 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="extract-content" Jan 20 09:35:39 crc kubenswrapper[4967]: E0120 09:35:39.831707 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="copy" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831718 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="copy" Jan 20 09:35:39 crc kubenswrapper[4967]: E0120 09:35:39.831745 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="gather" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831755 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="gather" Jan 20 09:35:39 crc kubenswrapper[4967]: E0120 09:35:39.831769 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="registry-server" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831779 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="registry-server" Jan 20 09:35:39 crc kubenswrapper[4967]: E0120 09:35:39.831797 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="extract-utilities" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831807 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="extract-utilities" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831981 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4baaa2ac-2a77-4886-9417-da415f5119ff" containerName="registry-server" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.831999 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="gather" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.832024 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86d9182-cb86-4c2d-87b0-c0e537e100e5" containerName="copy" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.833198 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.845112 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.971818 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.971950 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:39 crc kubenswrapper[4967]: I0120 09:35:39.971984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx6qn\" (UniqueName: \"kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.073304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.073405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.073427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx6qn\" (UniqueName: \"kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.074211 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.074238 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.097347 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx6qn\" (UniqueName: \"kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn\") pod \"community-operators-tqm8q\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.152057 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:40 crc kubenswrapper[4967]: I0120 09:35:40.480165 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:41 crc kubenswrapper[4967]: I0120 09:35:41.333956 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerID="36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab" exitCode=0 Jan 20 09:35:41 crc kubenswrapper[4967]: I0120 09:35:41.334036 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerDied","Data":"36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab"} Jan 20 09:35:41 crc kubenswrapper[4967]: I0120 09:35:41.334146 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerStarted","Data":"fc212ff39efdd061af9dbe9253ce8813e2eb43fa67c1d366e9be5f7e61c5d1f2"} Jan 20 09:35:42 crc kubenswrapper[4967]: I0120 09:35:42.345147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerStarted","Data":"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8"} Jan 20 09:35:43 crc kubenswrapper[4967]: I0120 09:35:43.352525 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerID="99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8" exitCode=0 Jan 20 09:35:43 crc kubenswrapper[4967]: I0120 09:35:43.352577 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerDied","Data":"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8"} Jan 20 09:35:44 crc kubenswrapper[4967]: I0120 09:35:44.360189 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerStarted","Data":"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568"} Jan 20 09:35:44 crc kubenswrapper[4967]: I0120 09:35:44.383971 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tqm8q" podStartSLOduration=2.580813927 podStartE2EDuration="5.383950181s" podCreationTimestamp="2026-01-20 09:35:39 +0000 UTC" firstStartedPulling="2026-01-20 09:35:41.335710967 +0000 UTC m=+2775.921011174" lastFinishedPulling="2026-01-20 09:35:44.138847221 +0000 UTC m=+2778.724147428" observedRunningTime="2026-01-20 09:35:44.379855498 +0000 UTC m=+2778.965155725" watchObservedRunningTime="2026-01-20 09:35:44.383950181 +0000 UTC m=+2778.969250388" Jan 20 09:35:48 crc kubenswrapper[4967]: I0120 09:35:48.474251 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:35:48 crc kubenswrapper[4967]: I0120 09:35:48.474574 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:35:48 crc kubenswrapper[4967]: I0120 09:35:48.474655 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:35:48 crc kubenswrapper[4967]: I0120 09:35:48.475350 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:35:48 crc kubenswrapper[4967]: I0120 09:35:48.475421 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c" gracePeriod=600 Jan 20 09:35:49 crc kubenswrapper[4967]: I0120 09:35:49.388784 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c" exitCode=0 Jan 20 09:35:49 crc kubenswrapper[4967]: I0120 09:35:49.388843 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c"} Jan 20 09:35:49 crc kubenswrapper[4967]: I0120 09:35:49.388894 4967 scope.go:117] "RemoveContainer" containerID="804c3ca5e35c5071d45528317fb4955f83656c70cfe68af3bf576d6405295723" Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.152892 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.153208 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.190679 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.398643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerStarted","Data":"a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a"} Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.437365 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:50 crc kubenswrapper[4967]: I0120 09:35:50.489136 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.408198 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tqm8q" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="registry-server" containerID="cri-o://3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568" gracePeriod=2 Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.784485 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.851196 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities\") pod \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.851304 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx6qn\" (UniqueName: \"kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn\") pod \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.851348 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content\") pod \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\" (UID: \"2a6a5255-009b-42a2-a1e3-c9a0a784c356\") " Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.852094 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities" (OuterVolumeSpecName: "utilities") pod "2a6a5255-009b-42a2-a1e3-c9a0a784c356" (UID: "2a6a5255-009b-42a2-a1e3-c9a0a784c356"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.858329 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn" (OuterVolumeSpecName: "kube-api-access-qx6qn") pod "2a6a5255-009b-42a2-a1e3-c9a0a784c356" (UID: "2a6a5255-009b-42a2-a1e3-c9a0a784c356"). InnerVolumeSpecName "kube-api-access-qx6qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.952876 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.952915 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx6qn\" (UniqueName: \"kubernetes.io/projected/2a6a5255-009b-42a2-a1e3-c9a0a784c356-kube-api-access-qx6qn\") on node \"crc\" DevicePath \"\"" Jan 20 09:35:52 crc kubenswrapper[4967]: I0120 09:35:52.975907 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a6a5255-009b-42a2-a1e3-c9a0a784c356" (UID: "2a6a5255-009b-42a2-a1e3-c9a0a784c356"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.054021 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a6a5255-009b-42a2-a1e3-c9a0a784c356-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.416710 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerID="3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568" exitCode=0 Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.416765 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerDied","Data":"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568"} Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.416800 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqm8q" event={"ID":"2a6a5255-009b-42a2-a1e3-c9a0a784c356","Type":"ContainerDied","Data":"fc212ff39efdd061af9dbe9253ce8813e2eb43fa67c1d366e9be5f7e61c5d1f2"} Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.416812 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqm8q" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.416825 4967 scope.go:117] "RemoveContainer" containerID="3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.436375 4967 scope.go:117] "RemoveContainer" containerID="99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.450034 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.453670 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tqm8q"] Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.459718 4967 scope.go:117] "RemoveContainer" containerID="36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.480391 4967 scope.go:117] "RemoveContainer" containerID="3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568" Jan 20 09:35:53 crc kubenswrapper[4967]: E0120 09:35:53.480980 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568\": container with ID starting with 3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568 not found: ID does not exist" containerID="3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.481033 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568"} err="failed to get container status \"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568\": rpc error: code = NotFound desc = could not find container \"3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568\": container with ID starting with 3cee721e9c5b2b0b14066fcee5bf26989ac5f82185677dc19fbd40a2ddd5f568 not found: ID does not exist" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.481117 4967 scope.go:117] "RemoveContainer" containerID="99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8" Jan 20 09:35:53 crc kubenswrapper[4967]: E0120 09:35:53.481369 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8\": container with ID starting with 99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8 not found: ID does not exist" containerID="99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.481406 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8"} err="failed to get container status \"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8\": rpc error: code = NotFound desc = could not find container \"99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8\": container with ID starting with 99dbf9089675ce05bb84c42ea4d9fae569fb937a1c51ac764d7120365927c7d8 not found: ID does not exist" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.481426 4967 scope.go:117] "RemoveContainer" containerID="36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab" Jan 20 09:35:53 crc kubenswrapper[4967]: E0120 09:35:53.481649 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab\": container with ID starting with 36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab not found: ID does not exist" containerID="36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.481711 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab"} err="failed to get container status \"36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab\": rpc error: code = NotFound desc = could not find container \"36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab\": container with ID starting with 36adf4661f55f2a944df44e968eed10d15e641f0b0945ded524c6cfa321637ab not found: ID does not exist" Jan 20 09:35:53 crc kubenswrapper[4967]: I0120 09:35:53.701629 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" path="/var/lib/kubelet/pods/2a6a5255-009b-42a2-a1e3-c9a0a784c356/volumes" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.757797 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:05 crc kubenswrapper[4967]: E0120 09:36:05.759021 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="extract-content" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.759037 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="extract-content" Jan 20 09:36:05 crc kubenswrapper[4967]: E0120 09:36:05.759048 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="extract-utilities" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.759056 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="extract-utilities" Jan 20 09:36:05 crc kubenswrapper[4967]: E0120 09:36:05.759074 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="registry-server" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.759083 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="registry-server" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.759232 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a6a5255-009b-42a2-a1e3-c9a0a784c356" containerName="registry-server" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.760278 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.773903 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.817143 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v7qw\" (UniqueName: \"kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.817355 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.817405 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.918233 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.918325 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v7qw\" (UniqueName: \"kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.918376 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.918730 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.918750 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:05 crc kubenswrapper[4967]: I0120 09:36:05.942936 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v7qw\" (UniqueName: \"kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw\") pod \"redhat-marketplace-vcrlj\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:06 crc kubenswrapper[4967]: I0120 09:36:06.080079 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:06 crc kubenswrapper[4967]: I0120 09:36:06.510884 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:07 crc kubenswrapper[4967]: I0120 09:36:07.497429 4967 generic.go:334] "Generic (PLEG): container finished" podID="8347bda7-747f-4d92-9044-43de65bf4f63" containerID="f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e" exitCode=0 Jan 20 09:36:07 crc kubenswrapper[4967]: I0120 09:36:07.497480 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerDied","Data":"f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e"} Jan 20 09:36:07 crc kubenswrapper[4967]: I0120 09:36:07.497781 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerStarted","Data":"c8f307b702c83e9562663843ea377e8a296dea1bc716ea11e680eadd201a68ac"} Jan 20 09:36:09 crc kubenswrapper[4967]: I0120 09:36:09.510482 4967 generic.go:334] "Generic (PLEG): container finished" podID="8347bda7-747f-4d92-9044-43de65bf4f63" containerID="27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e" exitCode=0 Jan 20 09:36:09 crc kubenswrapper[4967]: I0120 09:36:09.510567 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerDied","Data":"27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e"} Jan 20 09:36:10 crc kubenswrapper[4967]: I0120 09:36:10.519295 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerStarted","Data":"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958"} Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.080632 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.081229 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.136347 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.161918 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vcrlj" podStartSLOduration=8.335573664 podStartE2EDuration="11.161892934s" podCreationTimestamp="2026-01-20 09:36:05 +0000 UTC" firstStartedPulling="2026-01-20 09:36:07.499052713 +0000 UTC m=+2802.084352920" lastFinishedPulling="2026-01-20 09:36:10.325371983 +0000 UTC m=+2804.910672190" observedRunningTime="2026-01-20 09:36:10.53679374 +0000 UTC m=+2805.122093937" watchObservedRunningTime="2026-01-20 09:36:16.161892934 +0000 UTC m=+2810.747193141" Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.592794 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:16 crc kubenswrapper[4967]: I0120 09:36:16.644409 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:18 crc kubenswrapper[4967]: I0120 09:36:18.561352 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vcrlj" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="registry-server" containerID="cri-o://1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958" gracePeriod=2 Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.413648 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.493248 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities\") pod \"8347bda7-747f-4d92-9044-43de65bf4f63\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.493366 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content\") pod \"8347bda7-747f-4d92-9044-43de65bf4f63\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.493403 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v7qw\" (UniqueName: \"kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw\") pod \"8347bda7-747f-4d92-9044-43de65bf4f63\" (UID: \"8347bda7-747f-4d92-9044-43de65bf4f63\") " Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.494220 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities" (OuterVolumeSpecName: "utilities") pod "8347bda7-747f-4d92-9044-43de65bf4f63" (UID: "8347bda7-747f-4d92-9044-43de65bf4f63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.498522 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw" (OuterVolumeSpecName: "kube-api-access-5v7qw") pod "8347bda7-747f-4d92-9044-43de65bf4f63" (UID: "8347bda7-747f-4d92-9044-43de65bf4f63"). InnerVolumeSpecName "kube-api-access-5v7qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.522049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8347bda7-747f-4d92-9044-43de65bf4f63" (UID: "8347bda7-747f-4d92-9044-43de65bf4f63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.567366 4967 generic.go:334] "Generic (PLEG): container finished" podID="8347bda7-747f-4d92-9044-43de65bf4f63" containerID="1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958" exitCode=0 Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.567411 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerDied","Data":"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958"} Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.567435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vcrlj" event={"ID":"8347bda7-747f-4d92-9044-43de65bf4f63","Type":"ContainerDied","Data":"c8f307b702c83e9562663843ea377e8a296dea1bc716ea11e680eadd201a68ac"} Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.567453 4967 scope.go:117] "RemoveContainer" containerID="1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.567574 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vcrlj" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.587849 4967 scope.go:117] "RemoveContainer" containerID="27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.595501 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.595535 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8347bda7-747f-4d92-9044-43de65bf4f63-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.595548 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v7qw\" (UniqueName: \"kubernetes.io/projected/8347bda7-747f-4d92-9044-43de65bf4f63-kube-api-access-5v7qw\") on node \"crc\" DevicePath \"\"" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.601691 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.605359 4967 scope.go:117] "RemoveContainer" containerID="f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.606457 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vcrlj"] Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.622399 4967 scope.go:117] "RemoveContainer" containerID="1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958" Jan 20 09:36:19 crc kubenswrapper[4967]: E0120 09:36:19.622919 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958\": container with ID starting with 1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958 not found: ID does not exist" containerID="1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.622963 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958"} err="failed to get container status \"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958\": rpc error: code = NotFound desc = could not find container \"1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958\": container with ID starting with 1e6affaf4818b2c2e2059b3325e83d23b266b9bd0fdaec5fd1147dce86da5958 not found: ID does not exist" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.622991 4967 scope.go:117] "RemoveContainer" containerID="27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e" Jan 20 09:36:19 crc kubenswrapper[4967]: E0120 09:36:19.623444 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e\": container with ID starting with 27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e not found: ID does not exist" containerID="27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.623501 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e"} err="failed to get container status \"27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e\": rpc error: code = NotFound desc = could not find container \"27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e\": container with ID starting with 27654c9b67796c68c272199ced1ecc2b3b799c1cc9f6483c91bab9fe7518d95e not found: ID does not exist" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.623535 4967 scope.go:117] "RemoveContainer" containerID="f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e" Jan 20 09:36:19 crc kubenswrapper[4967]: E0120 09:36:19.623913 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e\": container with ID starting with f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e not found: ID does not exist" containerID="f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.623944 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e"} err="failed to get container status \"f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e\": rpc error: code = NotFound desc = could not find container \"f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e\": container with ID starting with f845237b38672db82d6e914148b7266ff3ee6d214099e4ad269160bad011712e not found: ID does not exist" Jan 20 09:36:19 crc kubenswrapper[4967]: I0120 09:36:19.700728 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" path="/var/lib/kubelet/pods/8347bda7-747f-4d92-9044-43de65bf4f63/volumes" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.253852 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:04 crc kubenswrapper[4967]: E0120 09:37:04.254519 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="extract-content" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.254530 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="extract-content" Jan 20 09:37:04 crc kubenswrapper[4967]: E0120 09:37:04.254543 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="extract-utilities" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.254551 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="extract-utilities" Jan 20 09:37:04 crc kubenswrapper[4967]: E0120 09:37:04.254567 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="registry-server" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.254573 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="registry-server" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.254705 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8347bda7-747f-4d92-9044-43de65bf4f63" containerName="registry-server" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.255546 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.271637 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.348270 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4skp9\" (UniqueName: \"kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.348363 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.348408 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.449134 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4skp9\" (UniqueName: \"kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.449239 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.449277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.450133 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.450202 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.469676 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4skp9\" (UniqueName: \"kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9\") pod \"redhat-operators-tqnjg\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.577025 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.828424 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:04 crc kubenswrapper[4967]: I0120 09:37:04.859255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerStarted","Data":"cb745e5b1ebd113d05b3d6772d24fd09e05f77f89b1c09338376312c8d3d73aa"} Jan 20 09:37:05 crc kubenswrapper[4967]: I0120 09:37:05.868295 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1a36dec-e817-4b10-a655-4c8d9742332c" containerID="a95c061b50e1fb2e69fc0e866eeb383cf02f4b0ab4b53648c95495ea41579181" exitCode=0 Jan 20 09:37:05 crc kubenswrapper[4967]: I0120 09:37:05.868336 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerDied","Data":"a95c061b50e1fb2e69fc0e866eeb383cf02f4b0ab4b53648c95495ea41579181"} Jan 20 09:37:06 crc kubenswrapper[4967]: I0120 09:37:06.876473 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerStarted","Data":"0f715141fd1143dc28bda6bc513e13876b730c34d456b86cf79417bdd259f980"} Jan 20 09:37:07 crc kubenswrapper[4967]: I0120 09:37:07.885668 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1a36dec-e817-4b10-a655-4c8d9742332c" containerID="0f715141fd1143dc28bda6bc513e13876b730c34d456b86cf79417bdd259f980" exitCode=0 Jan 20 09:37:07 crc kubenswrapper[4967]: I0120 09:37:07.885736 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerDied","Data":"0f715141fd1143dc28bda6bc513e13876b730c34d456b86cf79417bdd259f980"} Jan 20 09:37:08 crc kubenswrapper[4967]: I0120 09:37:08.894899 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerStarted","Data":"4e6438b7fd6957f797088739e35e2e21fc23f09f9b426a7f5459d31146678057"} Jan 20 09:37:08 crc kubenswrapper[4967]: I0120 09:37:08.916485 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tqnjg" podStartSLOduration=2.455597583 podStartE2EDuration="4.916464865s" podCreationTimestamp="2026-01-20 09:37:04 +0000 UTC" firstStartedPulling="2026-01-20 09:37:05.869593759 +0000 UTC m=+2860.454893966" lastFinishedPulling="2026-01-20 09:37:08.330461041 +0000 UTC m=+2862.915761248" observedRunningTime="2026-01-20 09:37:08.916206508 +0000 UTC m=+2863.501506735" watchObservedRunningTime="2026-01-20 09:37:08.916464865 +0000 UTC m=+2863.501765072" Jan 20 09:37:14 crc kubenswrapper[4967]: I0120 09:37:14.577242 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:14 crc kubenswrapper[4967]: I0120 09:37:14.577849 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:14 crc kubenswrapper[4967]: I0120 09:37:14.632255 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:14 crc kubenswrapper[4967]: I0120 09:37:14.966369 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:15 crc kubenswrapper[4967]: I0120 09:37:15.005857 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:16 crc kubenswrapper[4967]: I0120 09:37:16.943684 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tqnjg" podUID="d1a36dec-e817-4b10-a655-4c8d9742332c" containerName="registry-server" containerID="cri-o://4e6438b7fd6957f797088739e35e2e21fc23f09f9b426a7f5459d31146678057" gracePeriod=2 Jan 20 09:37:19 crc kubenswrapper[4967]: I0120 09:37:19.963241 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1a36dec-e817-4b10-a655-4c8d9742332c" containerID="4e6438b7fd6957f797088739e35e2e21fc23f09f9b426a7f5459d31146678057" exitCode=0 Jan 20 09:37:19 crc kubenswrapper[4967]: I0120 09:37:19.963338 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerDied","Data":"4e6438b7fd6957f797088739e35e2e21fc23f09f9b426a7f5459d31146678057"} Jan 20 09:37:19 crc kubenswrapper[4967]: I0120 09:37:19.963847 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnjg" event={"ID":"d1a36dec-e817-4b10-a655-4c8d9742332c","Type":"ContainerDied","Data":"cb745e5b1ebd113d05b3d6772d24fd09e05f77f89b1c09338376312c8d3d73aa"} Jan 20 09:37:19 crc kubenswrapper[4967]: I0120 09:37:19.963873 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb745e5b1ebd113d05b3d6772d24fd09e05f77f89b1c09338376312c8d3d73aa" Jan 20 09:37:19 crc kubenswrapper[4967]: I0120 09:37:19.970114 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.152820 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities\") pod \"d1a36dec-e817-4b10-a655-4c8d9742332c\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.152871 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content\") pod \"d1a36dec-e817-4b10-a655-4c8d9742332c\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.152970 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4skp9\" (UniqueName: \"kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9\") pod \"d1a36dec-e817-4b10-a655-4c8d9742332c\" (UID: \"d1a36dec-e817-4b10-a655-4c8d9742332c\") " Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.153740 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities" (OuterVolumeSpecName: "utilities") pod "d1a36dec-e817-4b10-a655-4c8d9742332c" (UID: "d1a36dec-e817-4b10-a655-4c8d9742332c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.165837 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9" (OuterVolumeSpecName: "kube-api-access-4skp9") pod "d1a36dec-e817-4b10-a655-4c8d9742332c" (UID: "d1a36dec-e817-4b10-a655-4c8d9742332c"). InnerVolumeSpecName "kube-api-access-4skp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.254102 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-utilities\") on node \"crc\" DevicePath \"\"" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.254135 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4skp9\" (UniqueName: \"kubernetes.io/projected/d1a36dec-e817-4b10-a655-4c8d9742332c-kube-api-access-4skp9\") on node \"crc\" DevicePath \"\"" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.281062 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1a36dec-e817-4b10-a655-4c8d9742332c" (UID: "d1a36dec-e817-4b10-a655-4c8d9742332c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.355365 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a36dec-e817-4b10-a655-4c8d9742332c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 20 09:37:20 crc kubenswrapper[4967]: I0120 09:37:20.969642 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnjg" Jan 20 09:37:21 crc kubenswrapper[4967]: I0120 09:37:21.002725 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:21 crc kubenswrapper[4967]: I0120 09:37:21.007070 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tqnjg"] Jan 20 09:37:21 crc kubenswrapper[4967]: I0120 09:37:21.702869 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1a36dec-e817-4b10-a655-4c8d9742332c" path="/var/lib/kubelet/pods/d1a36dec-e817-4b10-a655-4c8d9742332c/volumes" Jan 20 09:38:18 crc kubenswrapper[4967]: I0120 09:38:18.474700 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:38:18 crc kubenswrapper[4967]: I0120 09:38:18.475230 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:38:48 crc kubenswrapper[4967]: I0120 09:38:48.474974 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:38:48 crc kubenswrapper[4967]: I0120 09:38:48.475491 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.474633 4967 patch_prober.go:28] interesting pod/machine-config-daemon-bcgrn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.475386 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.475440 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.476121 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a"} pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.476181 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" containerName="machine-config-daemon" containerID="cri-o://a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a" gracePeriod=600 Jan 20 09:39:18 crc kubenswrapper[4967]: E0120 09:39:18.627804 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.696001 4967 generic.go:334] "Generic (PLEG): container finished" podID="5de94293-c4d1-4169-a8df-69843d57278a" containerID="a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a" exitCode=0 Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.696050 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" event={"ID":"5de94293-c4d1-4169-a8df-69843d57278a","Type":"ContainerDied","Data":"a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a"} Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.696085 4967 scope.go:117] "RemoveContainer" containerID="3339f7e3a811484dac7bd1a5f79e8b9ab302a37261ecfb7f75100f5d235eb02c" Jan 20 09:39:18 crc kubenswrapper[4967]: I0120 09:39:18.697254 4967 scope.go:117] "RemoveContainer" containerID="a9cab2cfb0690bc08dca816926b333e267c9ca7ff1bf7f2b92464a6c40aac87a" Jan 20 09:39:18 crc kubenswrapper[4967]: E0120 09:39:18.697726 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bcgrn_openshift-machine-config-operator(5de94293-c4d1-4169-a8df-69843d57278a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bcgrn" podUID="5de94293-c4d1-4169-a8df-69843d57278a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515133646532024455 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015133646532017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015133640323016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015133640323015455 5ustar corecore